2025-09-07T07:27:14.9551203Z Current runner version: '2.328.0' 2025-09-07T07:27:14.9558664Z Runner name: 'i-0c7ee175fda0aa6a4' 2025-09-07T07:27:14.9559696Z Runner group name: 'default' 2025-09-07T07:27:14.9560686Z Machine name: 'ip-10-0-72-124' 2025-09-07T07:27:14.9563958Z ##[group]GITHUB_TOKEN Permissions 2025-09-07T07:27:14.9566557Z Contents: read 2025-09-07T07:27:14.9567239Z Metadata: read 2025-09-07T07:27:14.9567924Z ##[endgroup] 2025-09-07T07:27:14.9570694Z Secret source: Actions 2025-09-07T07:27:14.9571601Z Prepare workflow directory 2025-09-07T07:27:15.0100853Z Prepare all required actions 2025-09-07T07:27:15.0141385Z Getting action download info 2025-09-07T07:27:15.3074485Z Download action repository 'pytorch/test-infra@main' (SHA:548a4bc624d43a01cdf165a63b041f0ae014ddbd) 2025-09-07T07:27:17.1964845Z Download action repository 'pytorch/pytorch@main' (SHA:93fb23d6fae7c4e82c4239a1033e522088742634) 2025-09-07T07:27:29.8746107Z Download action repository 'actions/setup-python@a26af69be951a213d495a4c3e4e4022e16d87065' (SHA:a26af69be951a213d495a4c3e4e4022e16d87065) 2025-09-07T07:27:30.1273566Z Download action repository 'aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722' (SHA:ececac1a45f3b08a01d2dd070d28d111c5fe6722) 2025-09-07T07:27:30.3487048Z Download action repository 'aws-actions/amazon-ecr-login@062b18b96a7aff071d4dc91bc00c4c1a7945b076' (SHA:062b18b96a7aff071d4dc91bc00c4c1a7945b076) 2025-09-07T07:27:30.5609058Z Download action repository 'seemethere/upload-artifact-s3@baba72d0712b404f646cebe0730933554ebce96a' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-09-07T07:27:30.8347684Z Getting action download info 2025-09-07T07:27:30.9582880Z Download action repository 'actions/checkout@v4' (SHA:08eba0b27e820071cde6df949e0beb9ba4906955) 2025-09-07T07:27:31.2209867Z Getting action download info 2025-09-07T07:27:31.3709485Z Download action repository 'nick-fields/retry@v3.0.0' (SHA:7152eba30c6575329ac0576536151aca5a72780e) 2025-09-07T07:27:31.5729388Z Getting action download info 2025-09-07T07:27:31.7307179Z Download action repository 'nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482' (SHA:3e91a01664abd3c5cd539100d10d33b9c5b68482) 2025-09-07T07:27:31.9330531Z Getting action download info 2025-09-07T07:27:32.0858623Z Uses: pytorch/pytorch/.github/workflows/_linux-test.yml@refs/heads/main (93fb23d6fae7c4e82c4239a1033e522088742634) 2025-09-07T07:27:32.0862573Z ##[group] Inputs 2025-09-07T07:27:32.0862921Z build-environment: linux-jammy-aarch64-py3.10 2025-09-07T07:27:32.0875863Z test-matrix: {"include": [{"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 1, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 2, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 3, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 4, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 5, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 6, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 7, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 8, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 9, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 1, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 2, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 3, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 4, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 5, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 6, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 7, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 8, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 9, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 10, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 11, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 12, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 13, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 14, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 15, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 1, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 2, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 3, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 4, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 5, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 6, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 7, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 8, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 9, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 10, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 11, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 12, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 13, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 14, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 15, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}]} 2025-09-07T07:27:32.0889179Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:27:32.0890007Z sync-tag: 2025-09-07T07:27:32.0890890Z timeout-minutes: 720 2025-09-07T07:27:32.0891163Z use-gha: 2025-09-07T07:27:32.0891670Z dashboard-tag: training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true 2025-09-07T07:27:32.0892257Z s3-bucket: gha-artifacts 2025-09-07T07:27:32.0892489Z aws-role-to-assume: 2025-09-07T07:27:32.0893062Z disable-monitor: false 2025-09-07T07:27:32.0893332Z monitor-log-interval: 15 2025-09-07T07:27:32.0893597Z monitor-data-collect-interval: 4 2025-09-07T07:27:32.0893880Z ##[endgroup] 2025-09-07T07:27:32.0894397Z Complete job name: linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal) 2025-09-07T07:27:32.1974458Z A job started hook has been configured by the self-hosted runner administrator 2025-09-07T07:27:32.2084549Z ##[group]Run '/home/ec2-user/runner-scripts/before_job.sh' 2025-09-07T07:27:32.2094756Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:27:32.2095452Z ##[endgroup] 2025-09-07T07:27:33.3911643Z Runner Type: linux.arm64.m7g.metal 2025-09-07T07:27:33.3912049Z Instance Type: m7g.metal 2025-09-07T07:27:33.3912281Z AMI Name: unknown 2025-09-07T07:27:33.3939077Z AMI ID: ami-022bbd2ccaf21691f 2025-09-07T07:27:38.2430994Z ##[group]Run pytorch/test-infra/.github/actions/setup-ssh@main 2025-09-07T07:27:38.2431419Z with: 2025-09-07T07:27:38.2431958Z github-secret: *** 2025-09-07T07:27:38.2432599Z instructions: All testing is done inside the container, to start an interactive session run: docker exec -it $(docker container ps --format '{{.ID}}') bash 2025-09-07T07:27:38.2433277Z activate-with-label: false 2025-09-07T07:27:38.2433514Z label: with-ssh 2025-09-07T07:27:38.2433722Z remove-existing-keys: true 2025-09-07T07:27:38.2433960Z fail-silently: true 2025-09-07T07:27:38.2434168Z env: 2025-09-07T07:27:38.2434345Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:27:38.2434586Z ##[endgroup] 2025-09-07T07:27:38.4447404Z Please see https://github.com/pytorch/pytorch/wiki/Debugging-using-with-ssh-for-Github-Actions for more info. 2025-09-07T07:27:38.4448965Z Not on pull request and ciflow reference could not be extracted, skipping adding ssh keys 2025-09-07T07:27:38.4614668Z ##[group]Run pytorch/pytorch/.github/actions/checkout-pytorch@main 2025-09-07T07:27:38.4615077Z with: 2025-09-07T07:27:38.4615262Z no-sudo: true 2025-09-07T07:27:38.4615463Z submodules: recursive 2025-09-07T07:27:38.4615695Z fetch-depth: 0 2025-09-07T07:27:38.4615880Z env: 2025-09-07T07:27:38.4616062Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:27:38.4616278Z ##[endgroup] 2025-09-07T07:27:38.4702496Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:27:38.4703357Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:27:38.4713258Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:27:38.4713603Z env: 2025-09-07T07:27:38.4713797Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:27:38.4714016Z ##[endgroup] 2025-09-07T07:27:38.4822265Z ##[group]Run # Use all available CPUs for fetching 2025-09-07T07:27:38.4822660Z # Use all available CPUs for fetching 2025-09-07T07:27:38.4822980Z cd "${GITHUB_WORKSPACE}" 2025-09-07T07:27:38.4823299Z git config --global fetch.parallel 0 2025-09-07T07:27:38.4823635Z git config --global submodule.fetchJobs 0 2025-09-07T07:27:38.4823927Z  2025-09-07T07:27:38.4824239Z # Clean workspace. The default checkout action should also do this, but 2025-09-07T07:27:38.4824647Z # do it here as well just in case 2025-09-07T07:27:38.4824923Z if [[ -d .git ]]; then 2025-09-07T07:27:38.4825176Z  if [ -z "${NO_SUDO}" ]; then 2025-09-07T07:27:38.4825458Z  sudo git clean -ffdx 2025-09-07T07:27:38.4825692Z  else 2025-09-07T07:27:38.4825889Z  git clean -ffdx 2025-09-07T07:27:38.4826112Z  fi 2025-09-07T07:27:38.4826289Z fi 2025-09-07T07:27:38.4832108Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:27:38.4832442Z env: 2025-09-07T07:27:38.4832700Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:27:38.4832937Z NO_SUDO: true 2025-09-07T07:27:38.4833121Z ##[endgroup] 2025-09-07T07:27:38.4986811Z ##[group]Run actions/checkout@v4 2025-09-07T07:27:38.4987069Z with: 2025-09-07T07:27:38.4987273Z ref: 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:27:38.4987547Z fetch-depth: 0 2025-09-07T07:27:38.4987746Z submodules: recursive 2025-09-07T07:27:38.4987960Z show-progress: false 2025-09-07T07:27:38.4988181Z repository: pytorch/pytorch 2025-09-07T07:27:38.4988546Z token: *** 2025-09-07T07:27:38.4988730Z ssh-strict: true 2025-09-07T07:27:38.4988919Z ssh-user: git 2025-09-07T07:27:38.4989387Z persist-credentials: true 2025-09-07T07:27:38.4989616Z clean: true 2025-09-07T07:27:38.4989832Z sparse-checkout-cone-mode: true 2025-09-07T07:27:38.4990081Z fetch-tags: false 2025-09-07T07:27:38.4990271Z lfs: false 2025-09-07T07:27:38.4990459Z set-safe-directory: true 2025-09-07T07:27:38.4990681Z env: 2025-09-07T07:27:38.4990856Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:27:38.4991069Z ##[endgroup] 2025-09-07T07:27:38.6441675Z Syncing repository: pytorch/pytorch 2025-09-07T07:27:38.6443025Z ##[group]Getting Git version info 2025-09-07T07:27:38.6443460Z Working directory is '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-09-07T07:27:38.6444038Z [command]/usr/bin/git version 2025-09-07T07:27:38.7814378Z git version 2.47.1 2025-09-07T07:27:38.7917990Z ##[endgroup] 2025-09-07T07:27:38.7927861Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/1038911c-7c59-4b9d-8fa0-9b18c03cd8da/.gitconfig' 2025-09-07T07:27:38.7978032Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/1038911c-7c59-4b9d-8fa0-9b18c03cd8da' before making global git config changes 2025-09-07T07:27:38.7978932Z Adding repository directory to the temporary git global config as a safe directory 2025-09-07T07:27:38.7983652Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-09-07T07:27:38.8077523Z Deleting the contents of '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-09-07T07:27:38.8080510Z ##[group]Initializing the repository 2025-09-07T07:27:38.8091870Z [command]/usr/bin/git init /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-09-07T07:27:38.8214519Z hint: Using 'master' as the name for the initial branch. This default branch name 2025-09-07T07:27:38.8215074Z hint: is subject to change. To configure the initial branch name to use in all 2025-09-07T07:27:38.8215574Z hint: of your new repositories, which will suppress this warning, call: 2025-09-07T07:27:38.8215934Z hint: 2025-09-07T07:27:38.8216204Z hint: git config --global init.defaultBranch 2025-09-07T07:27:38.8216505Z hint: 2025-09-07T07:27:38.8216789Z hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and 2025-09-07T07:27:38.8217295Z hint: 'development'. The just-created branch can be renamed via this command: 2025-09-07T07:27:38.8217671Z hint: 2025-09-07T07:27:38.8217857Z hint: git branch -m 2025-09-07T07:27:38.8316757Z Initialized empty Git repository in /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/ 2025-09-07T07:27:38.8326977Z [command]/usr/bin/git remote add origin https://github.com/pytorch/pytorch 2025-09-07T07:27:38.8388211Z ##[endgroup] 2025-09-07T07:27:38.8388611Z ##[group]Disabling automatic garbage collection 2025-09-07T07:27:38.8393391Z [command]/usr/bin/git config --local gc.auto 0 2025-09-07T07:27:38.8418502Z ##[endgroup] 2025-09-07T07:27:38.8418836Z ##[group]Setting up auth 2025-09-07T07:27:38.8425620Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-09-07T07:27:38.8451757Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-09-07T07:27:38.9235599Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-09-07T07:27:38.9260328Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-09-07T07:27:38.9724027Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-09-07T07:27:38.9811722Z ##[endgroup] 2025-09-07T07:27:38.9812127Z ##[group]Fetching the repository 2025-09-07T07:27:38.9818828Z [command]/usr/bin/git -c protocol.version=2 fetch --prune --no-recurse-submodules origin +refs/heads/*:refs/remotes/origin/* +refs/tags/*:refs/tags/* 2025-09-07T07:28:18.5388586Z From https://github.com/pytorch/pytorch 2025-09-07T07:28:18.5389015Z * [new branch] 160583 -> origin/160583 2025-09-07T07:28:18.5389573Z * [new branch] 2.6.0.dev20241004+ -> origin/2.6.0.dev20241004+ 2025-09-07T07:28:18.5390052Z * [new branch] 5addvllmbuild -> origin/5addvllmbuild 2025-09-07T07:28:18.5390592Z * [new branch] AaronWang04_addmmfusion_perftest -> origin/AaronWang04_addmmfusion_perftest 2025-09-07T07:28:18.5391210Z * [new branch] HDCharles-2.6.0-release-notes -> origin/HDCharles-2.6.0-release-notes 2025-09-07T07:28:18.5392098Z * [new branch] ISSUE-154849 -> origin/ISSUE-154849 2025-09-07T07:28:18.5394236Z * [new branch] JackCaoG/dynamo_make_fx_non_core_aten_ops -> origin/JackCaoG/dynamo_make_fx_non_core_aten_ops 2025-09-07T07:28:18.5395154Z * [new branch] NicoshevSVE128 -> origin/NicoshevSVE128 2025-09-07T07:28:18.5396153Z * [new branch] PR-AOTInductorNoneBug -> origin/PR-AOTInductorNoneBug 2025-09-07T07:28:18.5397219Z * [new branch] PR-AOTInductorNoneBugFix -> origin/PR-AOTInductorNoneBugFix 2025-09-07T07:28:18.5398469Z * [new branch] PR-FixConfigsIssue -> origin/PR-FixConfigsIssue 2025-09-07T07:28:18.5399013Z * [new branch] PR-NoneBugFix-viable -> origin/PR-NoneBugFix-viable 2025-09-07T07:28:18.5399912Z * [new branch] PR-ResetToZero -> origin/PR-ResetToZero 2025-09-07T07:28:18.5400928Z * [new branch] Update-Flash-Packaging -> origin/Update-Flash-Packaging 2025-09-07T07:28:18.5401814Z * [new branch] VLA_exp -> origin/VLA_exp 2025-09-07T07:28:18.5403548Z * [new branch] actually-run-mps-aot-inductor -> origin/actually-run-mps-aot-inductor 2025-09-07T07:28:18.5404919Z * [new branch] add-missing-args-normalization -> origin/add-missing-args-normalization 2025-09-07T07:28:18.5405895Z * [new branch] add-user-guide-structure -> origin/add-user-guide-structure 2025-09-07T07:28:18.5406935Z * [new branch] add-vllm-nightly-build -> origin/add-vllm-nightly-build 2025-09-07T07:28:18.5407853Z * [new branch] add_compile_benchmarking -> origin/add_compile_benchmarking 2025-09-07T07:28:18.5408793Z * [new branch] addmm-heuristic -> origin/addmm-heuristic 2025-09-07T07:28:18.5409703Z * [new branch] addsimde -> origin/addsimde 2025-09-07T07:28:18.5410682Z * [new branch] addvllmtest -> origin/addvllmtest 2025-09-07T07:28:18.5412043Z * [new branch] adi/acl_upgrade -> origin/adi/acl_upgrade 2025-09-07T07:28:18.5412985Z * [new branch] adi/test -> origin/adi/test 2025-09-07T07:28:18.5413899Z * [new branch] adi/test_bgemm -> origin/adi/test_bgemm 2025-09-07T07:28:18.5415001Z * [new branch] adi/test_fusions -> origin/adi/test_fusions 2025-09-07T07:28:18.5415820Z * [new branch] adi/test_onednn_v3.9 -> origin/adi/test_onednn_v3.9 2025-09-07T07:28:18.5416976Z * [new branch] adi/test_presve_change -> origin/adi/test_presve_change 2025-09-07T07:28:18.5417633Z * [new branch] adi/test_timm -> origin/adi/test_timm 2025-09-07T07:28:18.5418935Z * [new branch] adi/testpresve_change -> origin/adi/testpresve_change 2025-09-07T07:28:18.5420677Z * [new branch] aditew01/test/vec_bf16 -> origin/aditew01/test/vec_bf16 2025-09-07T07:28:18.5421675Z * [new branch] ah-globalfeedback-hook -> origin/ah-globalfeedback-hook 2025-09-07T07:28:18.5422585Z * [new branch] alt-disable -> origin/alt-disable 2025-09-07T07:28:18.5424105Z * [new branch] angelayi/aoti_additional_files -> origin/angelayi/aoti_additional_files 2025-09-07T07:28:18.5424931Z * [new branch] angelayi/aoti_inductor_fx -> origin/angelayi/aoti_inductor_fx 2025-09-07T07:28:18.5425788Z * [new branch] angelayi/benchmark -> origin/angelayi/benchmark 2025-09-07T07:28:18.5426850Z * [new branch] angelayi/benchmark2 -> origin/angelayi/benchmark2 2025-09-07T07:28:18.5427794Z * [new branch] angelayi/change_pytree_serialization -> origin/angelayi/change_pytree_serialization 2025-09-07T07:28:18.5428647Z * [new branch] angelayi/cpp_loader -> origin/angelayi/cpp_loader 2025-09-07T07:28:18.5429804Z * [new branch] angelayi/custom_op_subgraph -> origin/angelayi/custom_op_subgraph 2025-09-07T07:28:18.5430881Z * [new branch] angelayi/customop -> origin/angelayi/customop 2025-09-07T07:28:18.5432007Z * [new branch] angelayi/fake_cache_empty -> origin/angelayi/fake_cache_empty 2025-09-07T07:28:18.5433041Z * [new branch] angelayi/is_symbolic_tracing -> origin/angelayi/is_symbolic_tracing 2025-09-07T07:28:18.5433828Z * [new branch] angelayi/item -> origin/angelayi/item 2025-09-07T07:28:18.5434924Z * [new branch] angelayi/no_so_weight -> origin/angelayi/no_so_weight 2025-09-07T07:28:18.5435755Z * [new branch] angelayi/opoverload -> origin/angelayi/opoverload 2025-09-07T07:28:18.5436689Z * [new branch] angelayi/pattern -> origin/angelayi/pattern 2025-09-07T07:28:18.5437568Z * [new branch] angelayi/pytree -> origin/angelayi/pytree 2025-09-07T07:28:18.5438527Z * [new branch] angelayi/scan_layers -> origin/angelayi/scan_layers 2025-09-07T07:28:18.5439462Z * [new branch] angelayi/symint_input -> origin/angelayi/symint_input 2025-09-07T07:28:18.5440385Z * [new branch] angelayi/test_cpp -> origin/angelayi/test_cpp 2025-09-07T07:28:18.5441239Z * [new branch] angelayi/torch_size -> origin/angelayi/torch_size 2025-09-07T07:28:18.5442259Z * [new branch] aoti-cuda-alloc -> origin/aoti-cuda-alloc 2025-09-07T07:28:18.5443207Z * [new branch] aoti_target_windows -> origin/aoti_target_windows 2025-09-07T07:28:18.5444120Z * [new branch] aoti_weight_sharing -> origin/aoti_weight_sharing 2025-09-07T07:28:18.5445221Z * [new branch] atalman-inductor-perf-cu124 -> origin/atalman-inductor-perf-cu124 2025-09-07T07:28:18.5446144Z * [new branch] atalman-inductor-perf-cu124.1 -> origin/atalman-inductor-perf-cu124.1 2025-09-07T07:28:18.5447070Z * [new branch] atalman-patch-1 -> origin/atalman-patch-1 2025-09-07T07:28:18.5448085Z * [new branch] atalman-patch-3 -> origin/atalman-patch-3 2025-09-07T07:28:18.5448999Z * [new branch] atalman-patch-4 -> origin/atalman-patch-4 2025-09-07T07:28:18.5450050Z * [new branch] atalman-patch-5 -> origin/atalman-patch-5 2025-09-07T07:28:18.5451055Z * [new branch] atalman-patch-6 -> origin/atalman-patch-6 2025-09-07T07:28:18.5452074Z * [new branch] atalman_inductor_2.3.0 -> origin/atalman_inductor_2.3.0 2025-09-07T07:28:18.5452983Z * [new branch] atalman_inductor_2.3.1 -> origin/atalman_inductor_2.3.1 2025-09-07T07:28:18.5453887Z * [new branch] atalman_inductor_2.4.0 -> origin/atalman_inductor_2.4.0 2025-09-07T07:28:18.5454859Z * [new branch] atalman_inductor_2.4.x -> origin/atalman_inductor_2.4.x 2025-09-07T07:28:18.5455954Z * [new branch] autoupdate-transformers-pin-via-pr -> origin/autoupdate-transformers-pin-via-pr 2025-09-07T07:28:18.5457180Z * [new branch] bahuang/dtensor_demo -> origin/bahuang/dtensor_demo 2025-09-07T07:28:18.5458017Z * [new branch] bahuang/test -> origin/bahuang/test 2025-09-07T07:28:18.5459455Z * [new branch] base/1.5 -> origin/base/1.5 2025-09-07T07:28:18.5460538Z * [new branch] batching_sdpa_efficient_attention -> origin/batching_sdpa_efficient_attention 2025-09-07T07:28:18.5461385Z * [new branch] bc-lint-config -> origin/bc-lint-config 2025-09-07T07:28:18.5462381Z * [new branch] bc-lint-test-new-config -> origin/bc-lint-test-new-config 2025-09-07T07:28:18.5463402Z * [new branch] benchmark-updates -> origin/benchmark-updates 2025-09-07T07:28:18.5464406Z * [new branch] benchmarker_compat_with_do_bench -> origin/benchmarker_compat_with_do_bench 2025-09-07T07:28:18.5465416Z * [new branch] benchmarking-script -> origin/benchmarking-script 2025-09-07T07:28:18.5466917Z * [new branch] bertmaher/pinbump26 -> origin/bertmaher/pinbump26 2025-09-07T07:28:18.5468233Z * [new branch] bertrand/cutlass -> origin/bertrand/cutlass 2025-09-07T07:28:18.5469485Z * [new branch] bf/cg-custom-wrapper -> origin/bf/cg-custom-wrapper 2025-09-07T07:28:18.5470565Z * [new branch] bf/cg-or-error -> origin/bf/cg-or-error 2025-09-07T07:28:18.5471251Z * [new branch] bf/cg-remove-check -> origin/bf/cg-remove-check 2025-09-07T07:28:18.5472150Z * [new branch] bf/cg-skip-1-kernel -> origin/bf/cg-skip-1-kernel 2025-09-07T07:28:18.5472982Z * [new branch] bf/cudagraph -> origin/bf/cudagraph 2025-09-07T07:28:18.5474234Z * [new branch] bf/cudagraph-disable-input-mutation -> origin/bf/cudagraph-disable-input-mutation 2025-09-07T07:28:18.5475661Z * [new branch] bf/cudagraph-enable-input-mutation-support-benchmark -> origin/bf/cudagraph-enable-input-mutation-support-benchmark 2025-09-07T07:28:18.5476462Z * [new branch] bf/cudagraph-partition -> origin/bf/cudagraph-partition 2025-09-07T07:28:18.5477333Z * [new branch] bf/default-recompile-reason -> origin/bf/default-recompile-reason 2025-09-07T07:28:18.5478296Z * [new branch] bf/donated-buffer-bench -> origin/bf/donated-buffer-bench 2025-09-07T07:28:18.5479149Z * [new branch] bf/exp -> origin/bf/exp 2025-09-07T07:28:18.5479985Z * [new branch] bf/pa-non-divisible -> origin/bf/pa-non-divisible 2025-09-07T07:28:18.5480930Z * [new branch] bf/partition-move-cpu -> origin/bf/partition-move-cpu 2025-09-07T07:28:18.5481898Z * [new branch] bf/partition-turn-on -> origin/bf/partition-turn-on 2025-09-07T07:28:18.5482780Z * [new branch] bf/remove-check-55b0c39d -> origin/bf/remove-check-55b0c39d 2025-09-07T07:28:18.5483520Z * [new branch] bf/rope -> origin/bf/rope 2025-09-07T07:28:18.5484598Z * [new branch] bisect_perf_hf_T5_3acc6eac492 -> origin/bisect_perf_hf_T5_3acc6eac492 2025-09-07T07:28:18.5485500Z * [new branch] bisect_perf_hf_T5_3fcf66f61fb -> origin/bisect_perf_hf_T5_3fcf66f61fb 2025-09-07T07:28:18.5486317Z * [new branch] bisect_perf_hf_T5_4009d154129 -> origin/bisect_perf_hf_T5_4009d154129 2025-09-07T07:28:18.5487207Z * [new branch] bisect_perf_hf_T5_40d0740e73d -> origin/bisect_perf_hf_T5_40d0740e73d 2025-09-07T07:28:18.5488038Z * [new branch] bisect_perf_hf_T5_5268754e -> origin/bisect_perf_hf_T5_5268754e 2025-09-07T07:28:18.5488892Z * [new branch] bisect_perf_hf_T5_7d89a8d385c -> origin/bisect_perf_hf_T5_7d89a8d385c 2025-09-07T07:28:18.5489836Z * [new branch] bisect_perf_hf_T5_b7a25c1ee7c -> origin/bisect_perf_hf_T5_b7a25c1ee7c 2025-09-07T07:28:18.5490976Z * [new branch] bisect_perf_hf_T5_c25b201583f -> origin/bisect_perf_hf_T5_c25b201583f 2025-09-07T07:28:18.5491912Z * [new branch] bisect_perf_hf_T5_c93e57efac0 -> origin/bisect_perf_hf_T5_c93e57efac0 2025-09-07T07:28:18.5492840Z * [new branch] bisect_perf_hf_T5_ca9813ea149 -> origin/bisect_perf_hf_T5_ca9813ea149 2025-09-07T07:28:18.5493647Z * [new branch] bisect_perf_hf_T5_d65f194a -> origin/bisect_perf_hf_T5_d65f194a 2025-09-07T07:28:18.5494546Z * [new branch] bisect_perf_hf_T5_da94ab0b -> origin/bisect_perf_hf_T5_da94ab0b 2025-09-07T07:28:18.5495508Z * [new branch] bisect_perf_hf_T5_da94ab0b_new -> origin/bisect_perf_hf_T5_da94ab0b_new 2025-09-07T07:28:18.5496347Z * [new branch] bisect_perf_hf_T5_db4e8a1d8a8 -> origin/bisect_perf_hf_T5_db4e8a1d8a8 2025-09-07T07:28:18.5497241Z * [new branch] bisect_perf_hf_T5_e0d97e936a2 -> origin/bisect_perf_hf_T5_e0d97e936a2 2025-09-07T07:28:18.5498151Z * [new branch] bisect_perf_hf_T5_f23621ec563 -> origin/bisect_perf_hf_T5_f23621ec563 2025-09-07T07:28:18.5499513Z * [new branch] bowbao/bench_updates_stage -> origin/bowbao/bench_updates_stage 2025-09-07T07:28:18.5500376Z * [new branch] bowbao/dort_rewriter -> origin/bowbao/dort_rewriter 2025-09-07T07:28:18.5501293Z * [new branch] bowbao/wip_prs -> origin/bowbao/wip_prs 2025-09-07T07:28:18.5502456Z * [new branch] brister/break_tensorbox -> origin/brister/break_tensorbox 2025-09-07T07:28:18.5503376Z * [new branch] brister/custom_fx_backend -> origin/brister/custom_fx_backend 2025-09-07T07:28:18.5504270Z * [new branch] brister/fx_custom_triton -> origin/brister/fx_custom_triton 2025-09-07T07:28:18.5505063Z * [new branch] brister/tensor_box_output -> origin/brister/tensor_box_output 2025-09-07T07:28:18.5505988Z * [new branch] brister/tiled_reduction_no_numel_check -> origin/brister/tiled_reduction_no_numel_check 2025-09-07T07:28:18.5506928Z * [new branch] c57382a49 -> origin/c57382a49 2025-09-07T07:28:18.5507795Z * [new branch] ca_0431d47eaa -> origin/ca_0431d47eaa 2025-09-07T07:28:18.5508676Z * [new branch] ca_fix_0431d47eaa -> origin/ca_fix_0431d47eaa 2025-09-07T07:28:18.5510534Z * [new branch] camyll/revert-94bc900da97ad7f3c35b3b819bb53b23c74b581a-for-release-2.8 -> origin/camyll/revert-94bc900da97ad7f3c35b3b819bb53b23c74b581a-for-release-2.8 2025-09-07T07:28:18.5511575Z * [new branch] camyllh/test_setup_hooks_push -> origin/camyllh/test_setup_hooks_push 2025-09-07T07:28:18.5512575Z * [new branch] cherry-pick-149654-by-pytorch_bot_bot_ -> origin/cherry-pick-149654-by-pytorch_bot_bot_ 2025-09-07T07:28:18.5513597Z * [new branch] cherry-pick-151939-by-pytorch_bot_bot_ -> origin/cherry-pick-151939-by-pytorch_bot_bot_ 2025-09-07T07:28:18.5514611Z * [new branch] cherry-pick-154174-by-pytorch_bot_bot_ -> origin/cherry-pick-154174-by-pytorch_bot_bot_ 2025-09-07T07:28:18.5515601Z * [new branch] cherry-pick-156260-by-pytorch_bot_bot_ -> origin/cherry-pick-156260-by-pytorch_bot_bot_ 2025-09-07T07:28:18.5516642Z * [new branch] cherry-pick-157453-by-pytorch_bot_bot_ -> origin/cherry-pick-157453-by-pytorch_bot_bot_ 2025-09-07T07:28:18.5517565Z * [new branch] cherry-pick-157513-by-pytorch_bot_bot_ -> origin/cherry-pick-157513-by-pytorch_bot_bot_ 2025-09-07T07:28:18.5518472Z * [new branch] cherry-pick-157695-by-pytorch_bot_bot_ -> origin/cherry-pick-157695-by-pytorch_bot_bot_ 2025-09-07T07:28:18.5519400Z * [new branch] cherry-pick-157732-by-pytorch_bot_bot_ -> origin/cherry-pick-157732-by-pytorch_bot_bot_ 2025-09-07T07:28:18.5520357Z * [new branch] cherry-pick-158537-by-pytorch_bot_bot_ -> origin/cherry-pick-158537-by-pytorch_bot_bot_ 2025-09-07T07:28:18.5521264Z * [new branch] cherry-pick-159969-by-pytorch_bot_bot_ -> origin/cherry-pick-159969-by-pytorch_bot_bot_ 2025-09-07T07:28:18.5522272Z * [new branch] cherry-pick-160586-by-pytorch_bot_bot_ -> origin/cherry-pick-160586-by-pytorch_bot_bot_ 2025-09-07T07:28:18.5523484Z * [new branch] chilli/flex_vllm -> origin/chilli/flex_vllm 2025-09-07T07:28:18.5524545Z * [new branch] cleanup-inductor-benchmark-images -> origin/cleanup-inductor-benchmark-images 2025-09-07T07:28:18.5525315Z * [new branch] codex-testing -> origin/codex-testing 2025-09-07T07:28:18.5526926Z * [new branch] codex/add-helper-function-to-sizevars.py -> origin/codex/add-helper-function-to-sizevars.py 2025-09-07T07:28:18.5527877Z * [new branch] codex/add-helper-function-to-sizevars.py_2025-09-05 -> origin/codex/add-helper-function-to-sizevars.py_2025-09-05 2025-09-07T07:28:18.5528774Z * [new branch] codex/add-metadata-field-for-file-path -> origin/codex/add-metadata-field-for-file-path 2025-09-07T07:28:18.5529982Z * [new branch] codex/add-test-for-inductor-local-cache-behavior -> origin/codex/add-test-for-inductor-local-cache-behavior 2025-09-07T07:28:18.5531742Z * [new branch] codex/create-test-for-tensor-memory-leak-in-cudagraph -> origin/codex/create-test-for-tensor-memory-leak-in-cudagraph 2025-09-07T07:28:18.5532651Z * [new branch] codex/fix-issue-121219-in-pytorch -> origin/codex/fix-issue-121219-in-pytorch 2025-09-07T07:28:18.5533287Z * [new branch] codex/fix-issue-160415-in-pytorch -> origin/codex/fix-issue-160415-in-pytorch 2025-09-07T07:28:18.5534028Z * [new branch] codex/fix-noqengine-quantized-engine-support -> origin/codex/fix-noqengine-quantized-engine-support 2025-09-07T07:28:18.5534786Z * [new branch] codex/fix-pin_memory-error-handling -> origin/codex/fix-pin_memory-error-handling 2025-09-07T07:28:18.5535526Z * [new branch] codex/propose-fix-for-issue-160332 -> origin/codex/propose-fix-for-issue-160332 2025-09-07T07:28:18.5536491Z * [new branch] codex/refactor-lintrunner-config-to-use-uv-run -> origin/codex/refactor-lintrunner-config-to-use-uv-run 2025-09-07T07:28:18.5537436Z * [new branch] codex/remove-allow-untyped-defs-and-fix-type-errors -> origin/codex/remove-allow-untyped-defs-and-fix-type-errors 2025-09-07T07:28:18.5538263Z * [new branch] compile_fsdp2_disable_stream_and_event -> origin/compile_fsdp2_disable_stream_and_event 2025-09-07T07:28:18.5538994Z * [new branch] context_test -> origin/context_test 2025-09-07T07:28:18.5540333Z * [new branch] copilot/fix-157446 -> origin/copilot/fix-157446 2025-09-07T07:28:18.5541196Z * [new branch] copy_graph -> origin/copy_graph 2025-09-07T07:28:18.5542515Z * [new branch] cpio/fix_new_ami_tests -> origin/cpio/fix_new_ami_tests 2025-09-07T07:28:18.5543729Z * [new branch] csl/always_produce_xml -> origin/csl/always_produce_xml 2025-09-07T07:28:18.5544648Z * [new branch] csl/build_test_more_procs -> origin/csl/build_test_more_procs 2025-09-07T07:28:18.5545524Z * [new branch] csl/build_test_more_procs2 -> origin/csl/build_test_more_procs2 2025-09-07T07:28:18.5546374Z * [new branch] csl/disable_flaky_cpp_test -> origin/csl/disable_flaky_cpp_test 2025-09-07T07:28:18.5547206Z * [new branch] csl/disable_periodic_test -> origin/csl/disable_periodic_test 2025-09-07T07:28:18.5548256Z * [new branch] csl/exclude_rocm_viable_strict -> origin/csl/exclude_rocm_viable_strict 2025-09-07T07:28:18.5549359Z * [new branch] csl/katex -> origin/csl/katex 2025-09-07T07:28:18.5550275Z * [new branch] csl/larger_runner -> origin/csl/larger_runner 2025-09-07T07:28:18.5551168Z * [new branch] csl/lintrunner_stuff -> origin/csl/lintrunner_stuff 2025-09-07T07:28:18.5552085Z * [new branch] csl/mps_sharding -> origin/csl/mps_sharding 2025-09-07T07:28:18.5552897Z * [new branch] csl/multistage_docker -> origin/csl/multistage_docker 2025-09-07T07:28:18.5553793Z * [new branch] csl/name_link_check_job -> origin/csl/name_link_check_job 2025-09-07T07:28:18.5554663Z * [new branch] csl/no_keep_goin_rocm -> origin/csl/no_keep_goin_rocm 2025-09-07T07:28:18.5555574Z * [new branch] csl/not_600_timeout -> origin/csl/not_600_timeout 2025-09-07T07:28:18.5556428Z * [new branch] csl/revert_open -> origin/csl/revert_open 2025-09-07T07:28:18.5557347Z * [new branch] csl/skip_build -> origin/csl/skip_build 2025-09-07T07:28:18.5558270Z * [new branch] csl/test_cuda_build_large_runner -> origin/csl/test_cuda_build_large_runner 2025-09-07T07:28:18.5559119Z * [new branch] csl/win_sccache -> origin/csl/win_sccache 2025-09-07T07:28:18.5560098Z * [new branch] cublasltrelax2 -> origin/cublasltrelax2 2025-09-07T07:28:18.5560969Z * [new branch] cublasrelax2 -> origin/cublasrelax2 2025-09-07T07:28:18.5561958Z * [new branch] cudnnsdparefactor -> origin/cudnnsdparefactor 2025-09-07T07:28:18.5562726Z * [new branch] custom_lowering_dict -> origin/custom_lowering_dict 2025-09-07T07:28:18.5563924Z * [new branch] czhuge_muon_dev -> origin/czhuge_muon_dev 2025-09-07T07:28:18.5565375Z * [new branch] d4l3k/delete_hook -> origin/d4l3k/delete_hook 2025-09-07T07:28:18.5566410Z * [new branch] dcp_zoc -> origin/dcp_zoc 2025-09-07T07:28:18.5567370Z * [new branch] debug-guard -> origin/debug-guard 2025-09-07T07:28:18.5568314Z * [new branch] delete-quant-docs -> origin/delete-quant-docs 2025-09-07T07:28:18.5571666Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.2 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.2 2025-09-07T07:28:18.5572830Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.3 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.3 2025-09-07T07:28:18.5573974Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.4 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.4 2025-09-07T07:28:18.5575112Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.56.0 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.56.0 2025-09-07T07:28:18.5576072Z * [new branch] dependabot/pip/dot-ci/docker/protobuf-5.29.5 -> origin/dependabot/pip/dot-ci/docker/protobuf-5.29.5 2025-09-07T07:28:18.5577346Z * [new branch] dependabot/pip/dot-github/requirements/protobuf-5.29.5 -> origin/dependabot/pip/dot-github/requirements/protobuf-5.29.5 2025-09-07T07:28:18.5578561Z * [new branch] desertfire/test_cpp_wrapper -> origin/desertfire/test_cpp_wrapper 2025-09-07T07:28:18.5579447Z * [new branch] desertfire/triton-cpu-for-aarch64 -> origin/desertfire/triton-cpu-for-aarch64 2025-09-07T07:28:18.5581248Z * [new branch] dev/joona/MPSNDArrayAdd -> origin/dev/joona/MPSNDArrayAdd 2025-09-07T07:28:18.5582301Z * [new branch] dev/joona/Unranked -> origin/dev/joona/Unranked 2025-09-07T07:28:18.5583484Z * [new branch] dev/joona/cat -> origin/dev/joona/cat 2025-09-07T07:28:18.5584515Z * [new branch] dev/joona/cat_remove_graph -> origin/dev/joona/cat_remove_graph 2025-09-07T07:28:18.5585359Z * [new branch] dev/joona/embeddingbag -> origin/dev/joona/embeddingbag 2025-09-07T07:28:18.5586543Z * [new branch] dev/joona/getTensorsString -> origin/dev/joona/getTensorsString 2025-09-07T07:28:18.5587905Z * [new branch] dev/joona/maxpool2dwithindices_errmsg -> origin/dev/joona/maxpool2dwithindices_errmsg 2025-09-07T07:28:18.5589198Z * [new branch] dev/joona/mps_linear_macos14 -> origin/dev/joona/mps_linear_macos14 2025-09-07T07:28:18.5590508Z * [new branch] dev/joona/sdpa -> origin/dev/joona/sdpa 2025-09-07T07:28:18.5591757Z * [new branch] dev/joona/topk_newapi -> origin/dev/joona/topk_newapi 2025-09-07T07:28:18.5592879Z * [new branch] dev/joona/type_inf -> origin/dev/joona/type_inf 2025-09-07T07:28:18.5593925Z * [new branch] dev/joona/upsize3d -> origin/dev/joona/upsize3d 2025-09-07T07:28:18.5594857Z * [new branch] disable -> origin/disable 2025-09-07T07:28:18.5595831Z * [new branch] e2e-baseline -> origin/e2e-baseline 2025-09-07T07:28:18.5596711Z * [new branch] eigen_for_sparse_addmm_v2 -> origin/eigen_for_sparse_addmm_v2 2025-09-07T07:28:18.5598017Z * [new branch] embg/test_inductor_ci_128B -> origin/embg/test_inductor_ci_128B 2025-09-07T07:28:18.5599051Z * [new branch] embg/test_inductor_ci_base -> origin/embg/test_inductor_ci_base 2025-09-07T07:28:18.5599853Z * [new branch] embg/test_inductor_ci_control -> origin/embg/test_inductor_ci_control 2025-09-07T07:28:18.5600676Z * [new branch] embg/triton_l2_prefetch_128B -> origin/embg/triton_l2_prefetch_128B 2025-09-07T07:28:18.5601711Z * [new branch] embg/triton_l2_prefetch_256B -> origin/embg/triton_l2_prefetch_256B 2025-09-07T07:28:18.5602696Z * [new branch] eqy-patch-1 -> origin/eqy-patch-1 2025-09-07T07:28:18.5603645Z * [new branch] eqy-patch-2 -> origin/eqy-patch-2 2025-09-07T07:28:18.5604513Z * [new branch] eqy-patch-3 -> origin/eqy-patch-3 2025-09-07T07:28:18.5605395Z * [new branch] eqy-patch-4 -> origin/eqy-patch-4 2025-09-07T07:28:18.5606394Z * [new branch] example-convert-torch.nn -> origin/example-convert-torch.nn 2025-09-07T07:28:18.5607868Z * [new branch] exclamaforte/add-contiguous-threshold -> origin/exclamaforte/add-contiguous-threshold 2025-09-07T07:28:18.5608611Z * [new branch] exclamaforte/amd-ma -> origin/exclamaforte/amd-ma 2025-09-07T07:28:18.5609618Z * [new branch] exclamaforte/bump-transformer-version -> origin/exclamaforte/bump-transformer-version 2025-09-07T07:28:18.5610451Z * [new branch] exclamaforte/clear-feedback-savers -> origin/exclamaforte/clear-feedback-savers 2025-09-07T07:28:18.5611336Z * [new branch] exclamaforte/combo-kernels-perf-run -> origin/exclamaforte/combo-kernels-perf-run 2025-09-07T07:28:18.5612429Z * [new branch] exclamaforte/do_bench_refactor -> origin/exclamaforte/do_bench_refactor 2025-09-07T07:28:18.5613701Z * [new branch] exclamaforte/enable-mem-dep-fusion -> origin/exclamaforte/enable-mem-dep-fusion 2025-09-07T07:28:18.5614739Z * [new branch] exclamaforte/fix-exhaustive-autotuning -> origin/exclamaforte/fix-exhaustive-autotuning 2025-09-07T07:28:18.5615734Z * [new branch] exclamaforte/fix-exhuastive-autotuning-reland -> origin/exclamaforte/fix-exhuastive-autotuning-reland 2025-09-07T07:28:18.5616557Z * [new branch] exclamaforte/fix-trace-parsing-fx-svg -> origin/exclamaforte/fix-trace-parsing-fx-svg 2025-09-07T07:28:18.5617453Z * [new branch] exclamaforte/force-pointwise-cat-perf-run -> origin/exclamaforte/force-pointwise-cat-perf-run 2025-09-07T07:28:18.5618311Z * [new branch] exclamaforte/fusion-data -> origin/exclamaforte/fusion-data 2025-09-07T07:28:18.5619299Z * [new branch] exclamaforte/gemm-benchmark-run -> origin/exclamaforte/gemm-benchmark-run 2025-09-07T07:28:18.5620256Z * [new branch] exclamaforte/gemm-export-model -> origin/exclamaforte/gemm-export-model 2025-09-07T07:28:18.5621063Z * [new branch] exclamaforte/gemm-model -> origin/exclamaforte/gemm-model 2025-09-07T07:28:18.5622105Z * [new branch] exclamaforte/gemm-model-all-data-collection -> origin/exclamaforte/gemm-model-all-data-collection 2025-09-07T07:28:18.5622830Z * [new branch] exclamaforte/gemm-to-amd -> origin/exclamaforte/gemm-to-amd 2025-09-07T07:28:18.5623845Z * [new branch] exclamaforte/just-gemm-model -> origin/exclamaforte/just-gemm-model 2025-09-07T07:28:18.5624822Z * [new branch] exclamaforte/just-gemm-model-no-refactor -> origin/exclamaforte/just-gemm-model-no-refactor 2025-09-07T07:28:18.5625557Z * [new branch] exclamaforte/max-autotune-ieee -> origin/exclamaforte/max-autotune-ieee 2025-09-07T07:28:18.5626419Z * [new branch] exclamaforte/memory-counter -> origin/exclamaforte/memory-counter 2025-09-07T07:28:18.5627366Z * [new branch] exclamaforte/profile-diff-algo -> origin/exclamaforte/profile-diff-algo 2025-09-07T07:28:18.5628264Z * [new branch] exclamaforte/profiler-combo -> origin/exclamaforte/profiler-combo 2025-09-07T07:28:18.5629300Z * [new branch] exclamaforte/test_cpp_wrapper_mode -> origin/exclamaforte/test_cpp_wrapper_mode 2025-09-07T07:28:18.5630150Z * [new branch] exclamaforte/update-autotune-configs -> origin/exclamaforte/update-autotune-configs 2025-09-07T07:28:18.5631044Z * [new branch] exclamaforte/update-autotune-configs-2 -> origin/exclamaforte/update-autotune-configs-2 2025-09-07T07:28:18.5632269Z * [new branch] exclamforte/gemm-model-final -> origin/exclamforte/gemm-model-final 2025-09-07T07:28:18.5633173Z * [new branch] exec -> origin/exec 2025-09-07T07:28:18.5634095Z * [new branch] executorch-module-shim -> origin/executorch-module-shim 2025-09-07T07:28:18.5635104Z * [new branch] experimental-mosaic -> origin/experimental-mosaic 2025-09-07T07:28:18.5636101Z * [new branch] export-D58091437 -> origin/export-D58091437 2025-09-07T07:28:18.5637081Z * [new branch] export-D61047529 -> origin/export-D61047529 2025-09-07T07:28:18.5637988Z * [new branch] export-D70112642 -> origin/export-D70112642 2025-09-07T07:28:18.5639040Z * [new branch] export-D71412006 -> origin/export-D71412006 2025-09-07T07:28:18.5640082Z * [new branch] export-D73042989 -> origin/export-D73042989 2025-09-07T07:28:18.5641071Z * [new branch] export-D75183591 -> origin/export-D75183591 2025-09-07T07:28:18.5642014Z * [new branch] export-D75617432 -> origin/export-D75617432 2025-09-07T07:28:18.5642893Z * [new branch] export-D75659965 -> origin/export-D75659965 2025-09-07T07:28:18.5643823Z * [new branch] export-D76080931 -> origin/export-D76080931 2025-09-07T07:28:18.5644702Z * [new branch] export-D76797250 -> origin/export-D76797250 2025-09-07T07:28:18.5645601Z * [new branch] export-D76885271 -> origin/export-D76885271 2025-09-07T07:28:18.5646462Z * [new branch] export-D76885620 -> origin/export-D76885620 2025-09-07T07:28:18.5647350Z * [new branch] export-D76936623 -> origin/export-D76936623 2025-09-07T07:28:18.5648334Z * [new branch] export-D76958268 -> origin/export-D76958268 2025-09-07T07:28:18.5649249Z * [new branch] export-D78375400 -> origin/export-D78375400 2025-09-07T07:28:18.5650208Z * [new branch] export-D78431305 -> origin/export-D78431305 2025-09-07T07:28:18.5651455Z * [new branch] export-D78580107 -> origin/export-D78580107 2025-09-07T07:28:18.5652387Z * [new branch] export-D78822171 -> origin/export-D78822171 2025-09-07T07:28:18.5653300Z * [new branch] export-D78822351 -> origin/export-D78822351 2025-09-07T07:28:18.5654197Z * [new branch] export-D78822507 -> origin/export-D78822507 2025-09-07T07:28:18.5655078Z * [new branch] export-D78826994 -> origin/export-D78826994 2025-09-07T07:28:18.5656036Z * [new branch] export-D78894324 -> origin/export-D78894324 2025-09-07T07:28:18.5657247Z * [new branch] export-D78929245 -> origin/export-D78929245 2025-09-07T07:28:18.5658225Z * [new branch] export-D78934925 -> origin/export-D78934925 2025-09-07T07:28:18.5659160Z * [new branch] export-D78953203 -> origin/export-D78953203 2025-09-07T07:28:18.5660015Z * [new branch] export-D78953229 -> origin/export-D78953229 2025-09-07T07:28:18.5660867Z * [new branch] export-D78957093 -> origin/export-D78957093 2025-09-07T07:28:18.5661718Z * [new branch] export-D78957389 -> origin/export-D78957389 2025-09-07T07:28:18.5662715Z * [new branch] export-D78996107 -> origin/export-D78996107 2025-09-07T07:28:18.5663449Z * [new branch] export-D79026433 -> origin/export-D79026433 2025-09-07T07:28:18.5664427Z * [new branch] export-D79230339 -> origin/export-D79230339 2025-09-07T07:28:18.5665383Z * [new branch] export-D79319835 -> origin/export-D79319835 2025-09-07T07:28:18.5666279Z * [new branch] export-D79328456 -> origin/export-D79328456 2025-09-07T07:28:18.5667216Z * [new branch] export-D79534608 -> origin/export-D79534608 2025-09-07T07:28:18.5668379Z * [new branch] export-D79785974 -> origin/export-D79785974 2025-09-07T07:28:18.5669310Z * [new branch] export-D80025417 -> origin/export-D80025417 2025-09-07T07:28:18.5670227Z * [new branch] export-D80120333 -> origin/export-D80120333 2025-09-07T07:28:18.5671297Z * [new branch] export-D80214882 -> origin/export-D80214882 2025-09-07T07:28:18.5672184Z * [new branch] export-D80319069 -> origin/export-D80319069 2025-09-07T07:28:18.5673183Z * [new branch] export-D80321215 -> origin/export-D80321215 2025-09-07T07:28:18.5674018Z * [new branch] export-D80503451 -> origin/export-D80503451 2025-09-07T07:28:18.5674913Z * [new branch] export-D80771648 -> origin/export-D80771648 2025-09-07T07:28:18.5675757Z * [new branch] export-D80823877 -> origin/export-D80823877 2025-09-07T07:28:18.5676729Z * [new branch] export-D80948073 -> origin/export-D80948073 2025-09-07T07:28:18.5677768Z * [new branch] export-D80958642 -> origin/export-D80958642 2025-09-07T07:28:18.5678644Z * [new branch] export-D80970483 -> origin/export-D80970483 2025-09-07T07:28:18.5679570Z * [new branch] export-D81054193 -> origin/export-D81054193 2025-09-07T07:28:18.5680430Z * [new branch] export-D81060182 -> origin/export-D81060182 2025-09-07T07:28:18.5681446Z * [new branch] export-D81078973 -> origin/export-D81078973 2025-09-07T07:28:18.5682368Z * [new branch] export-D81204584 -> origin/export-D81204584 2025-09-07T07:28:18.5683371Z * [new branch] export-D81284190 -> origin/export-D81284190 2025-09-07T07:28:18.5684318Z * [new branch] export-D81299840 -> origin/export-D81299840 2025-09-07T07:28:18.5685174Z * [new branch] export-D81429090 -> origin/export-D81429090 2025-09-07T07:28:18.5686114Z * [new branch] export-D81698719 -> origin/export-D81698719 2025-09-07T07:28:18.5687094Z * [new branch] export-D81747409 -> origin/export-D81747409 2025-09-07T07:28:18.5688257Z * [new branch] exported-model-train-idempotent -> origin/exported-model-train-idempotent 2025-09-07T07:28:18.5689541Z * [new branch] ezyang/wip-aot-descriptors -> origin/ezyang/wip-aot-descriptors 2025-09-07T07:28:18.5690297Z * [new branch] fa_u8_brgemm -> origin/fa_u8_brgemm 2025-09-07T07:28:18.5691274Z * [new branch] fastmath_baseline -> origin/fastmath_baseline 2025-09-07T07:28:18.5692643Z * [new branch] fbcode/warm -> origin/fbcode/warm 2025-09-07T07:28:18.5693682Z * [new branch] fca -> origin/fca 2025-09-07T07:28:18.5694598Z * [new branch] fca2_ca5984c -> origin/fca2_ca5984c 2025-09-07T07:28:18.5695447Z * [new branch] fca5 -> origin/fca5 2025-09-07T07:28:18.5696827Z * [new branch] feature/function-numa-binding -> origin/feature/function-numa-binding 2025-09-07T07:28:18.5697872Z * [new branch] feature/function-numa-binding-take2 -> origin/feature/function-numa-binding-take2 2025-09-07T07:28:18.5698504Z * [new branch] feature/numa-nproc-fix -> origin/feature/numa-nproc-fix 2025-09-07T07:28:18.5699364Z * [new branch] feature/numa-signpost-serialize -> origin/feature/numa-signpost-serialize 2025-09-07T07:28:18.5700113Z * [new branch] feature/parallel-numa-binding -> origin/feature/parallel-numa-binding 2025-09-07T07:28:18.5701430Z * [new branch] fengyuan/external-proj -> origin/fengyuan/external-proj 2025-09-07T07:28:18.5702425Z * [new branch] fengyuan/out-of-tree-xpu-ops-improve-test -> origin/fengyuan/out-of-tree-xpu-ops-improve-test 2025-09-07T07:28:18.5703229Z * [new branch] fengyuan/out-of-tree-xpu-ops-remove-dtype -> origin/fengyuan/out-of-tree-xpu-ops-remove-dtype 2025-09-07T07:28:18.5703911Z * [new branch] fengyuan/test-xpu -> origin/fengyuan/test-xpu 2025-09-07T07:28:18.5705358Z * [new branch] ffast_math_baseline -> origin/ffast_math_baseline 2025-09-07T07:28:18.5706289Z * [new branch] ffast_math_target -> origin/ffast_math_target 2025-09-07T07:28:18.5707634Z * [new branch] findhao/base_commit -> origin/findhao/base_commit 2025-09-07T07:28:18.5708530Z * [new branch] findhao/base_commit1 -> origin/findhao/base_commit1 2025-09-07T07:28:18.5709386Z * [new branch] findhao/multistream2 -> origin/findhao/multistream2 2025-09-07T07:28:18.5710194Z * [new branch] findhao/multistream5 -> origin/findhao/multistream5 2025-09-07T07:28:18.5710992Z * [new branch] findhao/multistream6 -> origin/findhao/multistream6 2025-09-07T07:28:18.5711873Z * [new branch] findhao/operatorbench3 -> origin/findhao/operatorbench3 2025-09-07T07:28:18.5712727Z * [new branch] findhao/operatorbench5 -> origin/findhao/operatorbench5 2025-09-07T07:28:18.5713568Z * [new branch] findhao/tritonparse -> origin/findhao/tritonparse 2025-09-07T07:28:18.5714441Z * [new branch] fix -> origin/fix 2025-09-07T07:28:18.5715518Z * [new branch] fix-ck-gemm-template-format -> origin/fix-ck-gemm-template-format 2025-09-07T07:28:18.5716329Z * [new branch] fix-config-ignore -> origin/fix-config-ignore 2025-09-07T07:28:18.5717263Z * [new branch] fix-dict-guard -> origin/fix-dict-guard 2025-09-07T07:28:18.5718248Z * [new branch] fix-inductor-periodic-0528 -> origin/fix-inductor-periodic-0528 2025-09-07T07:28:18.5719071Z * [new branch] fix-mps-benchmark -> origin/fix-mps-benchmark 2025-09-07T07:28:18.5720068Z * [new branch] fix-rlease-feature-template -> origin/fix-rlease-feature-template 2025-09-07T07:28:18.5721038Z * [new branch] fix-run-condition-upload-results -> origin/fix-run-condition-upload-results 2025-09-07T07:28:18.5721804Z * [new branch] fix-torchbench -> origin/fix-torchbench 2025-09-07T07:28:18.5722726Z * [new branch] fix_153389 -> origin/fix_153389 2025-09-07T07:28:18.5723711Z * [new branch] fix_fsdp_rs_bucket2 -> origin/fix_fsdp_rs_bucket2 2025-09-07T07:28:18.5724647Z * [new branch] fix_inductor_peridic_tests -> origin/fix_inductor_peridic_tests 2025-09-07T07:28:18.5725452Z * [new branch] fix_ubn_159469 -> origin/fix_ubn_159469 2025-09-07T07:28:18.5726439Z * [new branch] fixes-triage -> origin/fixes-triage 2025-09-07T07:28:18.5727340Z * [new branch] fixflashinfer -> origin/fixflashinfer 2025-09-07T07:28:18.5728214Z * [new branch] flash_decoding_cpu -> origin/flash_decoding_cpu 2025-09-07T07:28:18.5729121Z * [new branch] flex-flash -> origin/flex-flash 2025-09-07T07:28:18.5730175Z * [new branch] flex-lowering -> origin/flex-lowering 2025-09-07T07:28:18.5731072Z * [new branch] flex-warning -> origin/flex-warning 2025-09-07T07:28:18.5732049Z * [new branch] flex_attention_functorch_grad -> origin/flex_attention_functorch_grad 2025-09-07T07:28:18.5733169Z * [new branch] flex_flash -> origin/flex_flash 2025-09-07T07:28:18.5734173Z * [new branch] flexdecode-gqa-groups -> origin/flexdecode-gqa-groups 2025-09-07T07:28:18.5735569Z * [new branch] fmassa/fix_memeff_sharding_rule -> origin/fmassa/fix_memeff_sharding_rule 2025-09-07T07:28:18.5736536Z * [new branch] fsdp2_trace_rules -> origin/fsdp2_trace_rules 2025-09-07T07:28:18.5737701Z * [new branch] fsdpv2_3d -> origin/fsdpv2_3d 2025-09-07T07:28:18.5738803Z * [new branch] fsdpv2_3d_m1 -> origin/fsdpv2_3d_m1 2025-09-07T07:28:18.5739786Z * [new branch] fx_cpp -> origin/fx_cpp 2025-09-07T07:28:18.5741092Z * [new branch] fy/fix-win -> origin/fy/fix-win 2025-09-07T07:28:18.5743479Z * [new branch] gh/AlnisM/1/base -> origin/gh/AlnisM/1/base 2025-09-07T07:28:18.5744305Z * [new branch] gh/AlnisM/1/head -> origin/gh/AlnisM/1/head 2025-09-07T07:28:18.5745751Z * [new branch] gh/CaoE/2/base -> origin/gh/CaoE/2/base 2025-09-07T07:28:18.5746625Z * [new branch] gh/CaoE/2/head -> origin/gh/CaoE/2/head 2025-09-07T07:28:18.5747532Z * [new branch] gh/CaoE/2/orig -> origin/gh/CaoE/2/orig 2025-09-07T07:28:18.5749088Z * [new branch] gh/ColinPeppler/79/base -> origin/gh/ColinPeppler/79/base 2025-09-07T07:28:18.5750076Z * [new branch] gh/ColinPeppler/79/head -> origin/gh/ColinPeppler/79/head 2025-09-07T07:28:18.5750954Z * [new branch] gh/ColinPeppler/79/orig -> origin/gh/ColinPeppler/79/orig 2025-09-07T07:28:18.5752326Z * [new branch] gh/ColinPeppler/80/base -> origin/gh/ColinPeppler/80/base 2025-09-07T07:28:18.5753376Z * [new branch] gh/ColinPeppler/80/head -> origin/gh/ColinPeppler/80/head 2025-09-07T07:28:18.5754234Z * [new branch] gh/ColinPeppler/80/orig -> origin/gh/ColinPeppler/80/orig 2025-09-07T07:28:18.5755843Z * [new branch] gh/EikanWang/67/base -> origin/gh/EikanWang/67/base 2025-09-07T07:28:18.5756688Z * [new branch] gh/EikanWang/67/head -> origin/gh/EikanWang/67/head 2025-09-07T07:28:18.5758009Z * [new branch] gh/EikanWang/80/base -> origin/gh/EikanWang/80/base 2025-09-07T07:28:18.5758883Z * [new branch] gh/EikanWang/80/head -> origin/gh/EikanWang/80/head 2025-09-07T07:28:18.5759782Z * [new branch] gh/EikanWang/80/orig -> origin/gh/EikanWang/80/orig 2025-09-07T07:28:18.5761066Z * [new branch] gh/EikanWang/81/base -> origin/gh/EikanWang/81/base 2025-09-07T07:28:18.5761889Z * [new branch] gh/EikanWang/81/head -> origin/gh/EikanWang/81/head 2025-09-07T07:28:18.5762737Z * [new branch] gh/EikanWang/81/orig -> origin/gh/EikanWang/81/orig 2025-09-07T07:28:18.5763941Z * [new branch] gh/EikanWang/82/base -> origin/gh/EikanWang/82/base 2025-09-07T07:28:18.5764784Z * [new branch] gh/EikanWang/82/head -> origin/gh/EikanWang/82/head 2025-09-07T07:28:18.5765750Z * [new branch] gh/EikanWang/82/orig -> origin/gh/EikanWang/82/orig 2025-09-07T07:28:18.5767719Z * [new branch] gh/Gasoonjia/1/base -> origin/gh/Gasoonjia/1/base 2025-09-07T07:28:18.5768594Z * [new branch] gh/Gasoonjia/1/head -> origin/gh/Gasoonjia/1/head 2025-09-07T07:28:18.5770426Z * [new branch] gh/H-Huang/131/base -> origin/gh/H-Huang/131/base 2025-09-07T07:28:18.5771077Z * [new branch] gh/H-Huang/131/head -> origin/gh/H-Huang/131/head 2025-09-07T07:28:18.5771974Z * [new branch] gh/H-Huang/131/orig -> origin/gh/H-Huang/131/orig 2025-09-07T07:28:18.5773204Z * [new branch] gh/H-Huang/132/base -> origin/gh/H-Huang/132/base 2025-09-07T07:28:18.5773997Z * [new branch] gh/H-Huang/132/head -> origin/gh/H-Huang/132/head 2025-09-07T07:28:18.5774890Z * [new branch] gh/H-Huang/132/orig -> origin/gh/H-Huang/132/orig 2025-09-07T07:28:18.5776215Z * [new branch] gh/H-Huang/180/base -> origin/gh/H-Huang/180/base 2025-09-07T07:28:18.5777036Z * [new branch] gh/H-Huang/180/head -> origin/gh/H-Huang/180/head 2025-09-07T07:28:18.5777956Z * [new branch] gh/H-Huang/180/orig -> origin/gh/H-Huang/180/orig 2025-09-07T07:28:18.5779113Z * [new branch] gh/H-Huang/182/base -> origin/gh/H-Huang/182/base 2025-09-07T07:28:18.5779926Z * [new branch] gh/H-Huang/182/head -> origin/gh/H-Huang/182/head 2025-09-07T07:28:18.5780780Z * [new branch] gh/H-Huang/182/orig -> origin/gh/H-Huang/182/orig 2025-09-07T07:28:18.5782153Z * [new branch] gh/H-Huang/187/base -> origin/gh/H-Huang/187/base 2025-09-07T07:28:18.5782924Z * [new branch] gh/H-Huang/187/head -> origin/gh/H-Huang/187/head 2025-09-07T07:28:18.5783772Z * [new branch] gh/H-Huang/187/orig -> origin/gh/H-Huang/187/orig 2025-09-07T07:28:18.5784924Z * [new branch] gh/H-Huang/202/base -> origin/gh/H-Huang/202/base 2025-09-07T07:28:18.5785873Z * [new branch] gh/H-Huang/202/head -> origin/gh/H-Huang/202/head 2025-09-07T07:28:18.5786721Z * [new branch] gh/H-Huang/202/orig -> origin/gh/H-Huang/202/orig 2025-09-07T07:28:18.5787883Z * [new branch] gh/H-Huang/203/base -> origin/gh/H-Huang/203/base 2025-09-07T07:28:18.5788725Z * [new branch] gh/H-Huang/203/head -> origin/gh/H-Huang/203/head 2025-09-07T07:28:18.5789572Z * [new branch] gh/H-Huang/203/orig -> origin/gh/H-Huang/203/orig 2025-09-07T07:28:18.5790805Z * [new branch] gh/H-Huang/204/base -> origin/gh/H-Huang/204/base 2025-09-07T07:28:18.5791643Z * [new branch] gh/H-Huang/204/head -> origin/gh/H-Huang/204/head 2025-09-07T07:28:18.5792540Z * [new branch] gh/H-Huang/204/orig -> origin/gh/H-Huang/204/orig 2025-09-07T07:28:18.5793740Z * [new branch] gh/H-Huang/205/base -> origin/gh/H-Huang/205/base 2025-09-07T07:28:18.5794631Z * [new branch] gh/H-Huang/205/head -> origin/gh/H-Huang/205/head 2025-09-07T07:28:18.5795487Z * [new branch] gh/H-Huang/205/orig -> origin/gh/H-Huang/205/orig 2025-09-07T07:28:18.5796653Z * [new branch] gh/H-Huang/206/base -> origin/gh/H-Huang/206/base 2025-09-07T07:28:18.5797528Z * [new branch] gh/H-Huang/206/head -> origin/gh/H-Huang/206/head 2025-09-07T07:28:18.5798375Z * [new branch] gh/H-Huang/206/orig -> origin/gh/H-Huang/206/orig 2025-09-07T07:28:18.5799573Z * [new branch] gh/H-Huang/207/base -> origin/gh/H-Huang/207/base 2025-09-07T07:28:18.5800425Z * [new branch] gh/H-Huang/207/head -> origin/gh/H-Huang/207/head 2025-09-07T07:28:18.5801235Z * [new branch] gh/H-Huang/207/orig -> origin/gh/H-Huang/207/orig 2025-09-07T07:28:18.5802448Z * [new branch] gh/H-Huang/208/base -> origin/gh/H-Huang/208/base 2025-09-07T07:28:18.5803320Z * [new branch] gh/H-Huang/208/head -> origin/gh/H-Huang/208/head 2025-09-07T07:28:18.5804334Z * [new branch] gh/H-Huang/208/orig -> origin/gh/H-Huang/208/orig 2025-09-07T07:28:18.5805332Z * [new branch] gh/H-Huang/209/base -> origin/gh/H-Huang/209/base 2025-09-07T07:28:18.5806212Z * [new branch] gh/H-Huang/209/head -> origin/gh/H-Huang/209/head 2025-09-07T07:28:18.5807011Z * [new branch] gh/H-Huang/209/orig -> origin/gh/H-Huang/209/orig 2025-09-07T07:28:18.5808266Z * [new branch] gh/H-Huang/210/base -> origin/gh/H-Huang/210/base 2025-09-07T07:28:18.5809089Z * [new branch] gh/H-Huang/210/head -> origin/gh/H-Huang/210/head 2025-09-07T07:28:18.5809944Z * [new branch] gh/H-Huang/210/orig -> origin/gh/H-Huang/210/orig 2025-09-07T07:28:18.5811103Z * [new branch] gh/H-Huang/211/base -> origin/gh/H-Huang/211/base 2025-09-07T07:28:18.5811957Z * [new branch] gh/H-Huang/211/head -> origin/gh/H-Huang/211/head 2025-09-07T07:28:18.5812798Z * [new branch] gh/H-Huang/211/orig -> origin/gh/H-Huang/211/orig 2025-09-07T07:28:18.5814013Z * [new branch] gh/H-Huang/212/base -> origin/gh/H-Huang/212/base 2025-09-07T07:28:18.5814834Z * [new branch] gh/H-Huang/212/head -> origin/gh/H-Huang/212/head 2025-09-07T07:28:18.5815633Z * [new branch] gh/H-Huang/212/orig -> origin/gh/H-Huang/212/orig 2025-09-07T07:28:18.5817431Z * [new branch] gh/H-Huang/213/base -> origin/gh/H-Huang/213/base 2025-09-07T07:28:18.5818303Z * [new branch] gh/H-Huang/213/head -> origin/gh/H-Huang/213/head 2025-09-07T07:28:18.5819185Z * [new branch] gh/H-Huang/213/orig -> origin/gh/H-Huang/213/orig 2025-09-07T07:28:18.5820384Z * [new branch] gh/H-Huang/214/base -> origin/gh/H-Huang/214/base 2025-09-07T07:28:18.5821221Z * [new branch] gh/H-Huang/214/head -> origin/gh/H-Huang/214/head 2025-09-07T07:28:18.5822078Z * [new branch] gh/H-Huang/214/orig -> origin/gh/H-Huang/214/orig 2025-09-07T07:28:18.5823593Z * [new branch] gh/IvanKobzarev/112/base -> origin/gh/IvanKobzarev/112/base 2025-09-07T07:28:18.5824521Z * [new branch] gh/IvanKobzarev/112/head -> origin/gh/IvanKobzarev/112/head 2025-09-07T07:28:18.5825388Z * [new branch] gh/IvanKobzarev/112/orig -> origin/gh/IvanKobzarev/112/orig 2025-09-07T07:28:18.5826656Z * [new branch] gh/IvanKobzarev/115/base -> origin/gh/IvanKobzarev/115/base 2025-09-07T07:28:18.5827562Z * [new branch] gh/IvanKobzarev/115/head -> origin/gh/IvanKobzarev/115/head 2025-09-07T07:28:18.5828440Z * [new branch] gh/IvanKobzarev/115/orig -> origin/gh/IvanKobzarev/115/orig 2025-09-07T07:28:18.5829949Z * [new branch] gh/IvanKobzarev/116/base -> origin/gh/IvanKobzarev/116/base 2025-09-07T07:28:18.5830896Z * [new branch] gh/IvanKobzarev/116/head -> origin/gh/IvanKobzarev/116/head 2025-09-07T07:28:18.5831788Z * [new branch] gh/IvanKobzarev/116/orig -> origin/gh/IvanKobzarev/116/orig 2025-09-07T07:28:18.5833104Z * [new branch] gh/IvanKobzarev/118/base -> origin/gh/IvanKobzarev/118/base 2025-09-07T07:28:18.5833971Z * [new branch] gh/IvanKobzarev/118/head -> origin/gh/IvanKobzarev/118/head 2025-09-07T07:28:18.5834837Z * [new branch] gh/IvanKobzarev/118/orig -> origin/gh/IvanKobzarev/118/orig 2025-09-07T07:28:18.5836220Z * [new branch] gh/IvanKobzarev/126/base -> origin/gh/IvanKobzarev/126/base 2025-09-07T07:28:18.5837144Z * [new branch] gh/IvanKobzarev/126/head -> origin/gh/IvanKobzarev/126/head 2025-09-07T07:28:18.5838047Z * [new branch] gh/IvanKobzarev/126/orig -> origin/gh/IvanKobzarev/126/orig 2025-09-07T07:28:18.5839348Z * [new branch] gh/IvanKobzarev/127/base -> origin/gh/IvanKobzarev/127/base 2025-09-07T07:28:18.5840361Z * [new branch] gh/IvanKobzarev/127/head -> origin/gh/IvanKobzarev/127/head 2025-09-07T07:28:18.5841112Z * [new branch] gh/IvanKobzarev/127/orig -> origin/gh/IvanKobzarev/127/orig 2025-09-07T07:28:18.5842353Z * [new branch] gh/IvanKobzarev/128/base -> origin/gh/IvanKobzarev/128/base 2025-09-07T07:28:18.5843250Z * [new branch] gh/IvanKobzarev/128/head -> origin/gh/IvanKobzarev/128/head 2025-09-07T07:28:18.5844104Z * [new branch] gh/IvanKobzarev/128/orig -> origin/gh/IvanKobzarev/128/orig 2025-09-07T07:28:18.5845379Z * [new branch] gh/IvanKobzarev/132/base -> origin/gh/IvanKobzarev/132/base 2025-09-07T07:28:18.5846295Z * [new branch] gh/IvanKobzarev/132/head -> origin/gh/IvanKobzarev/132/head 2025-09-07T07:28:18.5847153Z * [new branch] gh/IvanKobzarev/132/orig -> origin/gh/IvanKobzarev/132/orig 2025-09-07T07:28:18.5848849Z * [new branch] gh/IvanKobzarev/133/base -> origin/gh/IvanKobzarev/133/base 2025-09-07T07:28:18.5849936Z * [new branch] gh/IvanKobzarev/133/head -> origin/gh/IvanKobzarev/133/head 2025-09-07T07:28:18.5850895Z * [new branch] gh/IvanKobzarev/133/orig -> origin/gh/IvanKobzarev/133/orig 2025-09-07T07:28:18.5852039Z * [new branch] gh/IvanKobzarev/134/base -> origin/gh/IvanKobzarev/134/base 2025-09-07T07:28:18.5852901Z * [new branch] gh/IvanKobzarev/134/head -> origin/gh/IvanKobzarev/134/head 2025-09-07T07:28:18.5853722Z * [new branch] gh/IvanKobzarev/134/orig -> origin/gh/IvanKobzarev/134/orig 2025-09-07T07:28:18.5855234Z * [new branch] gh/IvanKobzarev/135/base -> origin/gh/IvanKobzarev/135/base 2025-09-07T07:28:18.5856088Z * [new branch] gh/IvanKobzarev/135/head -> origin/gh/IvanKobzarev/135/head 2025-09-07T07:28:18.5856935Z * [new branch] gh/IvanKobzarev/135/orig -> origin/gh/IvanKobzarev/135/orig 2025-09-07T07:28:18.5858275Z * [new branch] gh/IvanKobzarev/136/base -> origin/gh/IvanKobzarev/136/base 2025-09-07T07:28:18.5859317Z * [new branch] gh/IvanKobzarev/136/head -> origin/gh/IvanKobzarev/136/head 2025-09-07T07:28:18.5860129Z * [new branch] gh/IvanKobzarev/136/orig -> origin/gh/IvanKobzarev/136/orig 2025-09-07T07:28:18.5861116Z * [new branch] gh/IvanKobzarev/137/base -> origin/gh/IvanKobzarev/137/base 2025-09-07T07:28:18.5862034Z * [new branch] gh/IvanKobzarev/137/head -> origin/gh/IvanKobzarev/137/head 2025-09-07T07:28:18.5862848Z * [new branch] gh/IvanKobzarev/137/orig -> origin/gh/IvanKobzarev/137/orig 2025-09-07T07:28:18.5864094Z * [new branch] gh/IvanKobzarev/138/base -> origin/gh/IvanKobzarev/138/base 2025-09-07T07:28:18.5865223Z * [new branch] gh/IvanKobzarev/138/head -> origin/gh/IvanKobzarev/138/head 2025-09-07T07:28:18.5866277Z * [new branch] gh/IvanKobzarev/138/orig -> origin/gh/IvanKobzarev/138/orig 2025-09-07T07:28:18.5867541Z * [new branch] gh/IvanKobzarev/139/base -> origin/gh/IvanKobzarev/139/base 2025-09-07T07:28:18.5868363Z * [new branch] gh/IvanKobzarev/139/head -> origin/gh/IvanKobzarev/139/head 2025-09-07T07:28:18.5869208Z * [new branch] gh/IvanKobzarev/139/orig -> origin/gh/IvanKobzarev/139/orig 2025-09-07T07:28:18.5870551Z * [new branch] gh/IvanKobzarev/140/base -> origin/gh/IvanKobzarev/140/base 2025-09-07T07:28:18.5871354Z * [new branch] gh/IvanKobzarev/140/head -> origin/gh/IvanKobzarev/140/head 2025-09-07T07:28:18.5872263Z * [new branch] gh/IvanKobzarev/140/orig -> origin/gh/IvanKobzarev/140/orig 2025-09-07T07:28:18.5897214Z * [new branch] gh/IvanKobzarev/141/base -> origin/gh/IvanKobzarev/141/base 2025-09-07T07:28:18.5897914Z * [new branch] gh/IvanKobzarev/141/head -> origin/gh/IvanKobzarev/141/head 2025-09-07T07:28:18.5898677Z * [new branch] gh/IvanKobzarev/141/orig -> origin/gh/IvanKobzarev/141/orig 2025-09-07T07:28:18.5899209Z * [new branch] gh/IvanKobzarev/142/base -> origin/gh/IvanKobzarev/142/base 2025-09-07T07:28:18.5899722Z * [new branch] gh/IvanKobzarev/142/head -> origin/gh/IvanKobzarev/142/head 2025-09-07T07:28:18.5900230Z * [new branch] gh/IvanKobzarev/142/orig -> origin/gh/IvanKobzarev/142/orig 2025-09-07T07:28:18.5900734Z * [new branch] gh/IvanKobzarev/143/base -> origin/gh/IvanKobzarev/143/base 2025-09-07T07:28:18.5901240Z * [new branch] gh/IvanKobzarev/143/head -> origin/gh/IvanKobzarev/143/head 2025-09-07T07:28:18.5901757Z * [new branch] gh/IvanKobzarev/143/orig -> origin/gh/IvanKobzarev/143/orig 2025-09-07T07:28:18.5902261Z * [new branch] gh/IvanKobzarev/144/base -> origin/gh/IvanKobzarev/144/base 2025-09-07T07:28:18.5902773Z * [new branch] gh/IvanKobzarev/144/head -> origin/gh/IvanKobzarev/144/head 2025-09-07T07:28:18.5903280Z * [new branch] gh/IvanKobzarev/144/orig -> origin/gh/IvanKobzarev/144/orig 2025-09-07T07:28:18.5903787Z * [new branch] gh/IvanKobzarev/145/base -> origin/gh/IvanKobzarev/145/base 2025-09-07T07:28:18.5904290Z * [new branch] gh/IvanKobzarev/145/head -> origin/gh/IvanKobzarev/145/head 2025-09-07T07:28:18.5904798Z * [new branch] gh/IvanKobzarev/145/orig -> origin/gh/IvanKobzarev/145/orig 2025-09-07T07:28:18.5905306Z * [new branch] gh/IvanKobzarev/146/base -> origin/gh/IvanKobzarev/146/base 2025-09-07T07:28:18.5905824Z * [new branch] gh/IvanKobzarev/146/head -> origin/gh/IvanKobzarev/146/head 2025-09-07T07:28:18.5906337Z * [new branch] gh/IvanKobzarev/146/orig -> origin/gh/IvanKobzarev/146/orig 2025-09-07T07:28:18.5906851Z * [new branch] gh/NikhilAPatel/1/base -> origin/gh/NikhilAPatel/1/base 2025-09-07T07:28:18.5907365Z * [new branch] gh/NikhilAPatel/1/head -> origin/gh/NikhilAPatel/1/head 2025-09-07T07:28:18.5907865Z * [new branch] gh/NikhilAPatel/2/base -> origin/gh/NikhilAPatel/2/base 2025-09-07T07:28:18.5908361Z * [new branch] gh/NikhilAPatel/2/head -> origin/gh/NikhilAPatel/2/head 2025-09-07T07:28:18.5908859Z * [new branch] gh/NikhilAPatel/4/base -> origin/gh/NikhilAPatel/4/base 2025-09-07T07:28:18.5909366Z * [new branch] gh/NikhilAPatel/4/head -> origin/gh/NikhilAPatel/4/head 2025-09-07T07:28:18.5909850Z * [new branch] gh/PaliC/1/base -> origin/gh/PaliC/1/base 2025-09-07T07:28:18.5910473Z * [new branch] gh/PaliC/1/head -> origin/gh/PaliC/1/head 2025-09-07T07:28:18.5910909Z * [new branch] gh/PaliC/1/orig -> origin/gh/PaliC/1/orig 2025-09-07T07:28:18.5911353Z * [new branch] gh/PaliC/17/base -> origin/gh/PaliC/17/base 2025-09-07T07:28:18.5911796Z * [new branch] gh/PaliC/17/head -> origin/gh/PaliC/17/head 2025-09-07T07:28:18.5912234Z * [new branch] gh/PaliC/17/orig -> origin/gh/PaliC/17/orig 2025-09-07T07:28:18.5912669Z * [new branch] gh/PaliC/18/base -> origin/gh/PaliC/18/base 2025-09-07T07:28:18.5913103Z * [new branch] gh/PaliC/18/head -> origin/gh/PaliC/18/head 2025-09-07T07:28:18.5913541Z * [new branch] gh/PaliC/18/orig -> origin/gh/PaliC/18/orig 2025-09-07T07:28:18.5913980Z * [new branch] gh/PaliC/2/base -> origin/gh/PaliC/2/base 2025-09-07T07:28:18.5914431Z * [new branch] gh/PaliC/2/head -> origin/gh/PaliC/2/head 2025-09-07T07:28:18.5914870Z * [new branch] gh/PaliC/2/orig -> origin/gh/PaliC/2/orig 2025-09-07T07:28:18.5915304Z * [new branch] gh/PaliC/20/base -> origin/gh/PaliC/20/base 2025-09-07T07:28:18.5915874Z * [new branch] gh/PaliC/20/head -> origin/gh/PaliC/20/head 2025-09-07T07:28:18.5916331Z * [new branch] gh/PaliC/20/orig -> origin/gh/PaliC/20/orig 2025-09-07T07:28:18.5916778Z * [new branch] gh/PaliC/21/base -> origin/gh/PaliC/21/base 2025-09-07T07:28:18.5917219Z * [new branch] gh/PaliC/21/head -> origin/gh/PaliC/21/head 2025-09-07T07:28:18.5917654Z * [new branch] gh/PaliC/21/orig -> origin/gh/PaliC/21/orig 2025-09-07T07:28:18.5918098Z * [new branch] gh/PaliC/22/base -> origin/gh/PaliC/22/base 2025-09-07T07:28:18.5918540Z * [new branch] gh/PaliC/22/head -> origin/gh/PaliC/22/head 2025-09-07T07:28:18.5919596Z * [new branch] gh/PaliC/22/orig -> origin/gh/PaliC/22/orig 2025-09-07T07:28:18.5920427Z * [new branch] gh/PaliC/23/base -> origin/gh/PaliC/23/base 2025-09-07T07:28:18.5921257Z * [new branch] gh/PaliC/23/head -> origin/gh/PaliC/23/head 2025-09-07T07:28:18.5922117Z * [new branch] gh/PaliC/23/orig -> origin/gh/PaliC/23/orig 2025-09-07T07:28:18.5923304Z * [new branch] gh/PaliC/24/base -> origin/gh/PaliC/24/base 2025-09-07T07:28:18.5924115Z * [new branch] gh/PaliC/24/head -> origin/gh/PaliC/24/head 2025-09-07T07:28:18.5924943Z * [new branch] gh/PaliC/24/orig -> origin/gh/PaliC/24/orig 2025-09-07T07:28:18.5926477Z * [new branch] gh/PaulZhang12/17/base -> origin/gh/PaulZhang12/17/base 2025-09-07T07:28:18.5927314Z * [new branch] gh/PaulZhang12/17/head -> origin/gh/PaulZhang12/17/head 2025-09-07T07:28:18.5928724Z * [new branch] gh/PaulZhang12/20/base -> origin/gh/PaulZhang12/20/base 2025-09-07T07:28:18.5929609Z * [new branch] gh/PaulZhang12/20/head -> origin/gh/PaulZhang12/20/head 2025-09-07T07:28:18.5930524Z * [new branch] gh/PaulZhang12/20/orig -> origin/gh/PaulZhang12/20/orig 2025-09-07T07:28:18.5931778Z * [new branch] gh/PaulZhang12/21/base -> origin/gh/PaulZhang12/21/base 2025-09-07T07:28:18.5932715Z * [new branch] gh/PaulZhang12/21/head -> origin/gh/PaulZhang12/21/head 2025-09-07T07:28:18.5933532Z * [new branch] gh/PaulZhang12/21/orig -> origin/gh/PaulZhang12/21/orig 2025-09-07T07:28:18.5934786Z * [new branch] gh/PaulZhang12/22/base -> origin/gh/PaulZhang12/22/base 2025-09-07T07:28:18.5935609Z * [new branch] gh/PaulZhang12/22/head -> origin/gh/PaulZhang12/22/head 2025-09-07T07:28:18.5936558Z * [new branch] gh/PaulZhang12/22/orig -> origin/gh/PaulZhang12/22/orig 2025-09-07T07:28:18.5937779Z * [new branch] gh/PaulZhang12/23/base -> origin/gh/PaulZhang12/23/base 2025-09-07T07:28:18.5938624Z * [new branch] gh/PaulZhang12/23/head -> origin/gh/PaulZhang12/23/head 2025-09-07T07:28:18.5939487Z * [new branch] gh/PaulZhang12/23/orig -> origin/gh/PaulZhang12/23/orig 2025-09-07T07:28:18.5940522Z * [new branch] gh/PaulZhang12/24/base -> origin/gh/PaulZhang12/24/base 2025-09-07T07:28:18.5941395Z * [new branch] gh/PaulZhang12/24/head -> origin/gh/PaulZhang12/24/head 2025-09-07T07:28:18.5942372Z * [new branch] gh/PaulZhang12/24/orig -> origin/gh/PaulZhang12/24/orig 2025-09-07T07:28:18.5943500Z * [new branch] gh/PaulZhang12/25/base -> origin/gh/PaulZhang12/25/base 2025-09-07T07:28:18.5944364Z * [new branch] gh/PaulZhang12/25/head -> origin/gh/PaulZhang12/25/head 2025-09-07T07:28:18.5945308Z * [new branch] gh/PaulZhang12/25/orig -> origin/gh/PaulZhang12/25/orig 2025-09-07T07:28:18.5946788Z * [new branch] gh/SamGinzburg/11/base -> origin/gh/SamGinzburg/11/base 2025-09-07T07:28:18.5947742Z * [new branch] gh/SamGinzburg/11/head -> origin/gh/SamGinzburg/11/head 2025-09-07T07:28:18.5949570Z * [new branch] gh/Sidharth123-cpu/24/base -> origin/gh/Sidharth123-cpu/24/base 2025-09-07T07:28:18.5950709Z * [new branch] gh/Sidharth123-cpu/25/base -> origin/gh/Sidharth123-cpu/25/base 2025-09-07T07:28:18.5951771Z * [new branch] gh/Sidharth123-cpu/26/base -> origin/gh/Sidharth123-cpu/26/base 2025-09-07T07:28:18.5953036Z * [new branch] gh/Sidharth123-cpu/27/base -> origin/gh/Sidharth123-cpu/27/base 2025-09-07T07:28:18.5954517Z * [new branch] gh/StrongerXi/1/base -> origin/gh/StrongerXi/1/base 2025-09-07T07:28:18.5955378Z * [new branch] gh/StrongerXi/1/head -> origin/gh/StrongerXi/1/head 2025-09-07T07:28:18.5956592Z * [new branch] gh/StrongerXi/133/base -> origin/gh/StrongerXi/133/base 2025-09-07T07:28:18.5957463Z * [new branch] gh/StrongerXi/133/head -> origin/gh/StrongerXi/133/head 2025-09-07T07:28:18.5958402Z * [new branch] gh/StrongerXi/133/orig -> origin/gh/StrongerXi/133/orig 2025-09-07T07:28:18.5959507Z * [new branch] gh/StrongerXi/134/base -> origin/gh/StrongerXi/134/base 2025-09-07T07:28:18.5960408Z * [new branch] gh/StrongerXi/134/head -> origin/gh/StrongerXi/134/head 2025-09-07T07:28:18.5961274Z * [new branch] gh/StrongerXi/134/orig -> origin/gh/StrongerXi/134/orig 2025-09-07T07:28:18.5962549Z * [new branch] gh/StrongerXi/136/base -> origin/gh/StrongerXi/136/base 2025-09-07T07:28:18.5963409Z * [new branch] gh/StrongerXi/136/head -> origin/gh/StrongerXi/136/head 2025-09-07T07:28:18.5964310Z * [new branch] gh/StrongerXi/136/orig -> origin/gh/StrongerXi/136/orig 2025-09-07T07:28:18.5965460Z * [new branch] gh/StrongerXi/137/base -> origin/gh/StrongerXi/137/base 2025-09-07T07:28:18.5966353Z * [new branch] gh/StrongerXi/137/head -> origin/gh/StrongerXi/137/head 2025-09-07T07:28:18.5967245Z * [new branch] gh/StrongerXi/137/orig -> origin/gh/StrongerXi/137/orig 2025-09-07T07:28:18.5968371Z * [new branch] gh/StrongerXi/138/base -> origin/gh/StrongerXi/138/base 2025-09-07T07:28:18.5969240Z * [new branch] gh/StrongerXi/138/head -> origin/gh/StrongerXi/138/head 2025-09-07T07:28:18.5970130Z * [new branch] gh/StrongerXi/138/orig -> origin/gh/StrongerXi/138/orig 2025-09-07T07:28:18.5971273Z * [new branch] gh/StrongerXi/139/base -> origin/gh/StrongerXi/139/base 2025-09-07T07:28:18.5972130Z * [new branch] gh/StrongerXi/139/head -> origin/gh/StrongerXi/139/head 2025-09-07T07:28:18.5973090Z * [new branch] gh/StrongerXi/139/orig -> origin/gh/StrongerXi/139/orig 2025-09-07T07:28:18.5974304Z * [new branch] gh/StrongerXi/140/base -> origin/gh/StrongerXi/140/base 2025-09-07T07:28:18.5975120Z * [new branch] gh/StrongerXi/140/head -> origin/gh/StrongerXi/140/head 2025-09-07T07:28:18.5976106Z * [new branch] gh/StrongerXi/140/orig -> origin/gh/StrongerXi/140/orig 2025-09-07T07:28:18.5977229Z * [new branch] gh/StrongerXi/71/base -> origin/gh/StrongerXi/71/base 2025-09-07T07:28:18.5978050Z * [new branch] gh/StrongerXi/71/head -> origin/gh/StrongerXi/71/head 2025-09-07T07:28:18.5979178Z * [new branch] gh/StrongerXi/72/base -> origin/gh/StrongerXi/72/base 2025-09-07T07:28:18.5980084Z * [new branch] gh/StrongerXi/72/head -> origin/gh/StrongerXi/72/head 2025-09-07T07:28:18.5981622Z * [new branch] gh/XilunWu/133/base -> origin/gh/XilunWu/133/base 2025-09-07T07:28:18.5982454Z * [new branch] gh/XilunWu/133/head -> origin/gh/XilunWu/133/head 2025-09-07T07:28:18.5983574Z * [new branch] gh/XilunWu/133/orig -> origin/gh/XilunWu/133/orig 2025-09-07T07:28:18.5984575Z * [new branch] gh/XilunWu/139/base -> origin/gh/XilunWu/139/base 2025-09-07T07:28:18.5985544Z * [new branch] gh/XilunWu/139/head -> origin/gh/XilunWu/139/head 2025-09-07T07:28:18.5986301Z * [new branch] gh/XilunWu/139/orig -> origin/gh/XilunWu/139/orig 2025-09-07T07:28:18.5987559Z * [new branch] gh/XilunWu/143/base -> origin/gh/XilunWu/143/base 2025-09-07T07:28:18.5988535Z * [new branch] gh/XilunWu/143/head -> origin/gh/XilunWu/143/head 2025-09-07T07:28:18.5989340Z * [new branch] gh/XilunWu/143/orig -> origin/gh/XilunWu/143/orig 2025-09-07T07:28:18.5990674Z * [new branch] gh/XilunWu/144/base -> origin/gh/XilunWu/144/base 2025-09-07T07:28:18.5991560Z * [new branch] gh/XilunWu/144/head -> origin/gh/XilunWu/144/head 2025-09-07T07:28:18.5992404Z * [new branch] gh/XilunWu/144/orig -> origin/gh/XilunWu/144/orig 2025-09-07T07:28:18.5993621Z * [new branch] gh/XilunWu/145/base -> origin/gh/XilunWu/145/base 2025-09-07T07:28:18.5994469Z * [new branch] gh/XilunWu/145/head -> origin/gh/XilunWu/145/head 2025-09-07T07:28:18.5995380Z * [new branch] gh/XilunWu/145/orig -> origin/gh/XilunWu/145/orig 2025-09-07T07:28:18.5996479Z * [new branch] gh/XilunWu/146/base -> origin/gh/XilunWu/146/base 2025-09-07T07:28:18.5997340Z * [new branch] gh/XilunWu/146/head -> origin/gh/XilunWu/146/head 2025-09-07T07:28:18.5998228Z * [new branch] gh/XilunWu/146/orig -> origin/gh/XilunWu/146/orig 2025-09-07T07:28:18.5999438Z * [new branch] gh/XilunWu/147/base -> origin/gh/XilunWu/147/base 2025-09-07T07:28:18.6000389Z * [new branch] gh/XilunWu/147/head -> origin/gh/XilunWu/147/head 2025-09-07T07:28:18.6001279Z * [new branch] gh/XilunWu/147/orig -> origin/gh/XilunWu/147/orig 2025-09-07T07:28:18.6002305Z * [new branch] gh/XilunWu/148/base -> origin/gh/XilunWu/148/base 2025-09-07T07:28:18.6003176Z * [new branch] gh/XilunWu/148/head -> origin/gh/XilunWu/148/head 2025-09-07T07:28:18.6004038Z * [new branch] gh/XilunWu/148/orig -> origin/gh/XilunWu/148/orig 2025-09-07T07:28:18.6005125Z * [new branch] gh/XilunWu/149/base -> origin/gh/XilunWu/149/base 2025-09-07T07:28:18.6005951Z * [new branch] gh/XilunWu/149/head -> origin/gh/XilunWu/149/head 2025-09-07T07:28:18.6006804Z * [new branch] gh/XilunWu/149/orig -> origin/gh/XilunWu/149/orig 2025-09-07T07:28:18.6007861Z * [new branch] gh/XilunWu/150/base -> origin/gh/XilunWu/150/base 2025-09-07T07:28:18.6008804Z * [new branch] gh/XilunWu/150/head -> origin/gh/XilunWu/150/head 2025-09-07T07:28:18.6009706Z * [new branch] gh/XilunWu/150/orig -> origin/gh/XilunWu/150/orig 2025-09-07T07:28:18.6010993Z * [new branch] gh/XilunWu/151/base -> origin/gh/XilunWu/151/base 2025-09-07T07:28:18.6011852Z * [new branch] gh/XilunWu/151/head -> origin/gh/XilunWu/151/head 2025-09-07T07:28:18.6012781Z * [new branch] gh/XilunWu/151/orig -> origin/gh/XilunWu/151/orig 2025-09-07T07:28:18.6013904Z * [new branch] gh/XilunWu/152/base -> origin/gh/XilunWu/152/base 2025-09-07T07:28:18.6014643Z * [new branch] gh/XilunWu/152/head -> origin/gh/XilunWu/152/head 2025-09-07T07:28:18.6015425Z * [new branch] gh/XilunWu/152/orig -> origin/gh/XilunWu/152/orig 2025-09-07T07:28:18.6016795Z * [new branch] gh/XilunWu/153/base -> origin/gh/XilunWu/153/base 2025-09-07T07:28:18.6017833Z * [new branch] gh/XilunWu/153/head -> origin/gh/XilunWu/153/head 2025-09-07T07:28:18.6018622Z * [new branch] gh/XilunWu/153/orig -> origin/gh/XilunWu/153/orig 2025-09-07T07:28:18.6019960Z * [new branch] gh/XilunWu/160/base -> origin/gh/XilunWu/160/base 2025-09-07T07:28:18.6020801Z * [new branch] gh/XilunWu/160/head -> origin/gh/XilunWu/160/head 2025-09-07T07:28:18.6021759Z * [new branch] gh/XilunWu/160/orig -> origin/gh/XilunWu/160/orig 2025-09-07T07:28:18.6023028Z * [new branch] gh/XilunWu/161/base -> origin/gh/XilunWu/161/base 2025-09-07T07:28:18.6023819Z * [new branch] gh/XilunWu/161/head -> origin/gh/XilunWu/161/head 2025-09-07T07:28:18.6024681Z * [new branch] gh/XilunWu/161/orig -> origin/gh/XilunWu/161/orig 2025-09-07T07:28:18.6025930Z * [new branch] gh/XilunWu/163/base -> origin/gh/XilunWu/163/base 2025-09-07T07:28:18.6026801Z * [new branch] gh/XilunWu/163/head -> origin/gh/XilunWu/163/head 2025-09-07T07:28:18.6027690Z * [new branch] gh/XilunWu/163/orig -> origin/gh/XilunWu/163/orig 2025-09-07T07:28:18.6029296Z * [new branch] gh/XilunWu/164/base -> origin/gh/XilunWu/164/base 2025-09-07T07:28:18.6030269Z * [new branch] gh/XilunWu/164/head -> origin/gh/XilunWu/164/head 2025-09-07T07:28:18.6031236Z * [new branch] gh/XilunWu/164/orig -> origin/gh/XilunWu/164/orig 2025-09-07T07:28:18.6032581Z * [new branch] gh/XilunWu/165/base -> origin/gh/XilunWu/165/base 2025-09-07T07:28:18.6033657Z * [new branch] gh/XilunWu/165/head -> origin/gh/XilunWu/165/head 2025-09-07T07:28:18.6034563Z * [new branch] gh/XilunWu/165/orig -> origin/gh/XilunWu/165/orig 2025-09-07T07:28:18.6035927Z * [new branch] gh/XilunWu/166/base -> origin/gh/XilunWu/166/base 2025-09-07T07:28:18.6036824Z * [new branch] gh/XilunWu/166/head -> origin/gh/XilunWu/166/head 2025-09-07T07:28:18.6037694Z * [new branch] gh/XilunWu/166/orig -> origin/gh/XilunWu/166/orig 2025-09-07T07:28:18.6038880Z * [new branch] gh/XilunWu/167/base -> origin/gh/XilunWu/167/base 2025-09-07T07:28:18.6039765Z * [new branch] gh/XilunWu/167/head -> origin/gh/XilunWu/167/head 2025-09-07T07:28:18.6040684Z * [new branch] gh/XilunWu/167/orig -> origin/gh/XilunWu/167/orig 2025-09-07T07:28:18.6041956Z * [new branch] gh/XilunWu/168/base -> origin/gh/XilunWu/168/base 2025-09-07T07:28:18.6042781Z * [new branch] gh/XilunWu/168/head -> origin/gh/XilunWu/168/head 2025-09-07T07:28:18.6043610Z * [new branch] gh/XilunWu/168/orig -> origin/gh/XilunWu/168/orig 2025-09-07T07:28:18.6044833Z * [new branch] gh/XilunWu/169/base -> origin/gh/XilunWu/169/base 2025-09-07T07:28:18.6045777Z * [new branch] gh/XilunWu/169/head -> origin/gh/XilunWu/169/head 2025-09-07T07:28:18.6046657Z * [new branch] gh/XilunWu/169/orig -> origin/gh/XilunWu/169/orig 2025-09-07T07:28:18.6047733Z * [new branch] gh/XilunWu/170/base -> origin/gh/XilunWu/170/base 2025-09-07T07:28:18.6048614Z * [new branch] gh/XilunWu/170/head -> origin/gh/XilunWu/170/head 2025-09-07T07:28:18.6049422Z * [new branch] gh/XilunWu/170/orig -> origin/gh/XilunWu/170/orig 2025-09-07T07:28:18.6051013Z * [new branch] gh/XuehaiPan/14/base -> origin/gh/XuehaiPan/14/base 2025-09-07T07:28:18.6051883Z * [new branch] gh/XuehaiPan/14/head -> origin/gh/XuehaiPan/14/head 2025-09-07T07:28:18.6052662Z * [new branch] gh/XuehaiPan/14/orig -> origin/gh/XuehaiPan/14/orig 2025-09-07T07:28:18.6053987Z * [new branch] gh/XuehaiPan/179/base -> origin/gh/XuehaiPan/179/base 2025-09-07T07:28:18.6054963Z * [new branch] gh/XuehaiPan/179/head -> origin/gh/XuehaiPan/179/head 2025-09-07T07:28:18.6055822Z * [new branch] gh/XuehaiPan/179/orig -> origin/gh/XuehaiPan/179/orig 2025-09-07T07:28:18.6057119Z * [new branch] gh/XuehaiPan/189/base -> origin/gh/XuehaiPan/189/base 2025-09-07T07:28:18.6057939Z * [new branch] gh/XuehaiPan/189/head -> origin/gh/XuehaiPan/189/head 2025-09-07T07:28:18.6058858Z * [new branch] gh/XuehaiPan/189/orig -> origin/gh/XuehaiPan/189/orig 2025-09-07T07:28:18.6060014Z * [new branch] gh/XuehaiPan/232/base -> origin/gh/XuehaiPan/232/base 2025-09-07T07:28:18.6060860Z * [new branch] gh/XuehaiPan/232/head -> origin/gh/XuehaiPan/232/head 2025-09-07T07:28:18.6061750Z * [new branch] gh/XuehaiPan/232/orig -> origin/gh/XuehaiPan/232/orig 2025-09-07T07:28:18.6062965Z * [new branch] gh/XuehaiPan/249/base -> origin/gh/XuehaiPan/249/base 2025-09-07T07:28:18.6063889Z * [new branch] gh/XuehaiPan/249/head -> origin/gh/XuehaiPan/249/head 2025-09-07T07:28:18.6064736Z * [new branch] gh/XuehaiPan/249/orig -> origin/gh/XuehaiPan/249/orig 2025-09-07T07:28:18.6065931Z * [new branch] gh/XuehaiPan/253/base -> origin/gh/XuehaiPan/253/base 2025-09-07T07:28:18.6066785Z * [new branch] gh/XuehaiPan/253/head -> origin/gh/XuehaiPan/253/head 2025-09-07T07:28:18.6067596Z * [new branch] gh/XuehaiPan/253/orig -> origin/gh/XuehaiPan/253/orig 2025-09-07T07:28:18.6068775Z * [new branch] gh/XuehaiPan/254/base -> origin/gh/XuehaiPan/254/base 2025-09-07T07:28:18.6069625Z * [new branch] gh/XuehaiPan/254/head -> origin/gh/XuehaiPan/254/head 2025-09-07T07:28:18.6070628Z * [new branch] gh/XuehaiPan/254/orig -> origin/gh/XuehaiPan/254/orig 2025-09-07T07:28:18.6071742Z * [new branch] gh/XuehaiPan/255/base -> origin/gh/XuehaiPan/255/base 2025-09-07T07:28:18.6072605Z * [new branch] gh/XuehaiPan/255/head -> origin/gh/XuehaiPan/255/head 2025-09-07T07:28:18.6073392Z * [new branch] gh/XuehaiPan/255/orig -> origin/gh/XuehaiPan/255/orig 2025-09-07T07:28:18.6074617Z * [new branch] gh/XuehaiPan/257/base -> origin/gh/XuehaiPan/257/base 2025-09-07T07:28:18.6075469Z * [new branch] gh/XuehaiPan/257/head -> origin/gh/XuehaiPan/257/head 2025-09-07T07:28:18.6076292Z * [new branch] gh/XuehaiPan/257/orig -> origin/gh/XuehaiPan/257/orig 2025-09-07T07:28:18.6077468Z * [new branch] gh/XuehaiPan/271/base -> origin/gh/XuehaiPan/271/base 2025-09-07T07:28:18.6078336Z * [new branch] gh/XuehaiPan/271/head -> origin/gh/XuehaiPan/271/head 2025-09-07T07:28:18.6079215Z * [new branch] gh/XuehaiPan/271/orig -> origin/gh/XuehaiPan/271/orig 2025-09-07T07:28:18.6080448Z * [new branch] gh/XuehaiPan/290/base -> origin/gh/XuehaiPan/290/base 2025-09-07T07:28:18.6081426Z * [new branch] gh/XuehaiPan/290/head -> origin/gh/XuehaiPan/290/head 2025-09-07T07:28:18.6082150Z * [new branch] gh/XuehaiPan/290/orig -> origin/gh/XuehaiPan/290/orig 2025-09-07T07:28:18.6083357Z * [new branch] gh/XuehaiPan/343/base -> origin/gh/XuehaiPan/343/base 2025-09-07T07:28:18.6093174Z * [new branch] gh/XuehaiPan/343/head -> origin/gh/XuehaiPan/343/head 2025-09-07T07:28:18.6093661Z * [new branch] gh/XuehaiPan/343/orig -> origin/gh/XuehaiPan/343/orig 2025-09-07T07:28:18.6094146Z * [new branch] gh/XuehaiPan/347/base -> origin/gh/XuehaiPan/347/base 2025-09-07T07:28:18.6094631Z * [new branch] gh/XuehaiPan/347/head -> origin/gh/XuehaiPan/347/head 2025-09-07T07:28:18.6095269Z * [new branch] gh/XuehaiPan/347/orig -> origin/gh/XuehaiPan/347/orig 2025-09-07T07:28:18.6095764Z * [new branch] gh/XuehaiPan/348/base -> origin/gh/XuehaiPan/348/base 2025-09-07T07:28:18.6096245Z * [new branch] gh/XuehaiPan/348/head -> origin/gh/XuehaiPan/348/head 2025-09-07T07:28:18.6096726Z * [new branch] gh/XuehaiPan/348/orig -> origin/gh/XuehaiPan/348/orig 2025-09-07T07:28:18.6097203Z * [new branch] gh/XuehaiPan/350/base -> origin/gh/XuehaiPan/350/base 2025-09-07T07:28:18.6097678Z * [new branch] gh/XuehaiPan/350/head -> origin/gh/XuehaiPan/350/head 2025-09-07T07:28:18.6098157Z * [new branch] gh/XuehaiPan/350/orig -> origin/gh/XuehaiPan/350/orig 2025-09-07T07:28:18.6098646Z * [new branch] gh/XuehaiPan/356/base -> origin/gh/XuehaiPan/356/base 2025-09-07T07:28:18.6099133Z * [new branch] gh/XuehaiPan/356/head -> origin/gh/XuehaiPan/356/head 2025-09-07T07:28:18.6099617Z * [new branch] gh/XuehaiPan/356/orig -> origin/gh/XuehaiPan/356/orig 2025-09-07T07:28:18.6100094Z * [new branch] gh/XuehaiPan/357/base -> origin/gh/XuehaiPan/357/base 2025-09-07T07:28:18.6100574Z * [new branch] gh/XuehaiPan/357/head -> origin/gh/XuehaiPan/357/head 2025-09-07T07:28:18.6101061Z * [new branch] gh/XuehaiPan/357/orig -> origin/gh/XuehaiPan/357/orig 2025-09-07T07:28:18.6101545Z * [new branch] gh/XuehaiPan/358/base -> origin/gh/XuehaiPan/358/base 2025-09-07T07:28:18.6102029Z * [new branch] gh/XuehaiPan/358/head -> origin/gh/XuehaiPan/358/head 2025-09-07T07:28:18.6102864Z * [new branch] gh/XuehaiPan/358/orig -> origin/gh/XuehaiPan/358/orig 2025-09-07T07:28:18.6104175Z * [new branch] gh/XuehaiPan/359/base -> origin/gh/XuehaiPan/359/base 2025-09-07T07:28:18.6104902Z * [new branch] gh/XuehaiPan/359/head -> origin/gh/XuehaiPan/359/head 2025-09-07T07:28:18.6105771Z * [new branch] gh/XuehaiPan/359/orig -> origin/gh/XuehaiPan/359/orig 2025-09-07T07:28:18.6106910Z * [new branch] gh/XuehaiPan/360/base -> origin/gh/XuehaiPan/360/base 2025-09-07T07:28:18.6107841Z * [new branch] gh/XuehaiPan/360/head -> origin/gh/XuehaiPan/360/head 2025-09-07T07:28:18.6108663Z * [new branch] gh/XuehaiPan/360/orig -> origin/gh/XuehaiPan/360/orig 2025-09-07T07:28:18.6109936Z * [new branch] gh/XuehaiPan/365/base -> origin/gh/XuehaiPan/365/base 2025-09-07T07:28:18.6110759Z * [new branch] gh/XuehaiPan/365/head -> origin/gh/XuehaiPan/365/head 2025-09-07T07:28:18.6111591Z * [new branch] gh/XuehaiPan/365/orig -> origin/gh/XuehaiPan/365/orig 2025-09-07T07:28:18.6112796Z * [new branch] gh/XuehaiPan/366/base -> origin/gh/XuehaiPan/366/base 2025-09-07T07:28:18.6113680Z * [new branch] gh/XuehaiPan/366/head -> origin/gh/XuehaiPan/366/head 2025-09-07T07:28:18.6114975Z * [new branch] gh/XuehaiPan/369/base -> origin/gh/XuehaiPan/369/base 2025-09-07T07:28:18.6115836Z * [new branch] gh/XuehaiPan/369/head -> origin/gh/XuehaiPan/369/head 2025-09-07T07:28:18.6116685Z * [new branch] gh/XuehaiPan/369/orig -> origin/gh/XuehaiPan/369/orig 2025-09-07T07:28:18.6117858Z * [new branch] gh/XuehaiPan/370/base -> origin/gh/XuehaiPan/370/base 2025-09-07T07:28:18.6118745Z * [new branch] gh/XuehaiPan/370/head -> origin/gh/XuehaiPan/370/head 2025-09-07T07:28:18.6119613Z * [new branch] gh/XuehaiPan/370/orig -> origin/gh/XuehaiPan/370/orig 2025-09-07T07:28:18.6120779Z * [new branch] gh/XuehaiPan/380/base -> origin/gh/XuehaiPan/380/base 2025-09-07T07:28:18.6121637Z * [new branch] gh/XuehaiPan/380/head -> origin/gh/XuehaiPan/380/head 2025-09-07T07:28:18.6122598Z * [new branch] gh/XuehaiPan/380/orig -> origin/gh/XuehaiPan/380/orig 2025-09-07T07:28:18.6123722Z * [new branch] gh/XuehaiPan/381/base -> origin/gh/XuehaiPan/381/base 2025-09-07T07:28:18.6124655Z * [new branch] gh/XuehaiPan/381/head -> origin/gh/XuehaiPan/381/head 2025-09-07T07:28:18.6125923Z * [new branch] gh/XuehaiPan/382/base -> origin/gh/XuehaiPan/382/base 2025-09-07T07:28:18.6126747Z * [new branch] gh/XuehaiPan/382/head -> origin/gh/XuehaiPan/382/head 2025-09-07T07:28:18.6127643Z * [new branch] gh/XuehaiPan/382/orig -> origin/gh/XuehaiPan/382/orig 2025-09-07T07:28:18.6128976Z * [new branch] gh/XuehaiPan/383/base -> origin/gh/XuehaiPan/383/base 2025-09-07T07:28:18.6129794Z * [new branch] gh/XuehaiPan/383/head -> origin/gh/XuehaiPan/383/head 2025-09-07T07:28:18.6130642Z * [new branch] gh/XuehaiPan/383/orig -> origin/gh/XuehaiPan/383/orig 2025-09-07T07:28:18.6131834Z * [new branch] gh/XuehaiPan/384/base -> origin/gh/XuehaiPan/384/base 2025-09-07T07:28:18.6132750Z * [new branch] gh/XuehaiPan/384/head -> origin/gh/XuehaiPan/384/head 2025-09-07T07:28:18.6133571Z * [new branch] gh/XuehaiPan/384/orig -> origin/gh/XuehaiPan/384/orig 2025-09-07T07:28:18.6134862Z * [new branch] gh/XuehaiPan/385/base -> origin/gh/XuehaiPan/385/base 2025-09-07T07:28:18.6135688Z * [new branch] gh/XuehaiPan/385/head -> origin/gh/XuehaiPan/385/head 2025-09-07T07:28:18.6136498Z * [new branch] gh/XuehaiPan/385/orig -> origin/gh/XuehaiPan/385/orig 2025-09-07T07:28:18.6137619Z * [new branch] gh/XuehaiPan/386/base -> origin/gh/XuehaiPan/386/base 2025-09-07T07:28:18.6138493Z * [new branch] gh/XuehaiPan/386/head -> origin/gh/XuehaiPan/386/head 2025-09-07T07:28:18.6139404Z * [new branch] gh/XuehaiPan/386/orig -> origin/gh/XuehaiPan/386/orig 2025-09-07T07:28:18.6140599Z * [new branch] gh/XuehaiPan/387/base -> origin/gh/XuehaiPan/387/base 2025-09-07T07:28:18.6141406Z * [new branch] gh/XuehaiPan/387/head -> origin/gh/XuehaiPan/387/head 2025-09-07T07:28:18.6142306Z * [new branch] gh/XuehaiPan/387/orig -> origin/gh/XuehaiPan/387/orig 2025-09-07T07:28:18.6143805Z * [new branch] gh/ZainRizvi/1/base -> origin/gh/ZainRizvi/1/base 2025-09-07T07:28:18.6144696Z * [new branch] gh/ZainRizvi/1/head -> origin/gh/ZainRizvi/1/head 2025-09-07T07:28:18.6145755Z * [new branch] gh/ZainRizvi/2/base -> origin/gh/ZainRizvi/2/base 2025-09-07T07:28:18.6146612Z * [new branch] gh/ZainRizvi/2/head -> origin/gh/ZainRizvi/2/head 2025-09-07T07:28:18.6147750Z * [new branch] gh/ZainRizvi/3/base -> origin/gh/ZainRizvi/3/base 2025-09-07T07:28:18.6148486Z * [new branch] gh/ZainRizvi/3/head -> origin/gh/ZainRizvi/3/head 2025-09-07T07:28:18.6149674Z * [new branch] gh/ZainRizvi/4/base -> origin/gh/ZainRizvi/4/base 2025-09-07T07:28:18.6150581Z * [new branch] gh/ZainRizvi/4/head -> origin/gh/ZainRizvi/4/head 2025-09-07T07:28:18.6151642Z * [new branch] gh/ZainRizvi/5/base -> origin/gh/ZainRizvi/5/base 2025-09-07T07:28:18.6152374Z * [new branch] gh/ZainRizvi/5/head -> origin/gh/ZainRizvi/5/head 2025-09-07T07:28:18.6153499Z * [new branch] gh/ZainRizvi/6/base -> origin/gh/ZainRizvi/6/base 2025-09-07T07:28:18.6154379Z * [new branch] gh/ZainRizvi/6/head -> origin/gh/ZainRizvi/6/head 2025-09-07T07:28:18.6155227Z * [new branch] gh/ZainRizvi/6/orig -> origin/gh/ZainRizvi/6/orig 2025-09-07T07:28:18.6156406Z * [new branch] gh/ZainRizvi/7/base -> origin/gh/ZainRizvi/7/base 2025-09-07T07:28:18.6157407Z * [new branch] gh/ZainRizvi/7/head -> origin/gh/ZainRizvi/7/head 2025-09-07T07:28:18.6158092Z * [new branch] gh/ZainRizvi/7/orig -> origin/gh/ZainRizvi/7/orig 2025-09-07T07:28:18.6159371Z * [new branch] gh/ZainRizvi/8/base -> origin/gh/ZainRizvi/8/base 2025-09-07T07:28:18.6160241Z * [new branch] gh/ZainRizvi/8/head -> origin/gh/ZainRizvi/8/head 2025-09-07T07:28:18.6161463Z * [new branch] gh/ZainRizvi/9/base -> origin/gh/ZainRizvi/9/base 2025-09-07T07:28:18.6162306Z * [new branch] gh/ZainRizvi/9/head -> origin/gh/ZainRizvi/9/head 2025-09-07T07:28:18.6163157Z * [new branch] gh/ZainRizvi/9/orig -> origin/gh/ZainRizvi/9/orig 2025-09-07T07:28:18.6164648Z * [new branch] gh/ZhiweiYan-96/39/base -> origin/gh/ZhiweiYan-96/39/base 2025-09-07T07:28:18.6165662Z * [new branch] gh/ZhiweiYan-96/39/head -> origin/gh/ZhiweiYan-96/39/head 2025-09-07T07:28:18.6166622Z * [new branch] gh/ZhiweiYan-96/39/orig -> origin/gh/ZhiweiYan-96/39/orig 2025-09-07T07:28:18.6167826Z * [new branch] gh/ZhiweiYan-96/44/base -> origin/gh/ZhiweiYan-96/44/base 2025-09-07T07:28:18.6168668Z * [new branch] gh/ZhiweiYan-96/44/head -> origin/gh/ZhiweiYan-96/44/head 2025-09-07T07:28:18.6169717Z * [new branch] gh/ZhiweiYan-96/45/base -> origin/gh/ZhiweiYan-96/45/base 2025-09-07T07:28:18.6170535Z * [new branch] gh/ZhiweiYan-96/45/head -> origin/gh/ZhiweiYan-96/45/head 2025-09-07T07:28:18.6171825Z * [new branch] gh/ZhiweiYan-96/49/base -> origin/gh/ZhiweiYan-96/49/base 2025-09-07T07:28:18.6172644Z * [new branch] gh/ZhiweiYan-96/49/head -> origin/gh/ZhiweiYan-96/49/head 2025-09-07T07:28:18.6173830Z * [new branch] gh/ZhiweiYan-96/62/base -> origin/gh/ZhiweiYan-96/62/base 2025-09-07T07:28:18.6174625Z * [new branch] gh/ZhiweiYan-96/62/head -> origin/gh/ZhiweiYan-96/62/head 2025-09-07T07:28:18.6175873Z * [new branch] gh/ZhiweiYan-96/64/base -> origin/gh/ZhiweiYan-96/64/base 2025-09-07T07:28:18.6176746Z * [new branch] gh/ZhiweiYan-96/64/head -> origin/gh/ZhiweiYan-96/64/head 2025-09-07T07:28:18.6177599Z * [new branch] gh/ZhiweiYan-96/64/orig -> origin/gh/ZhiweiYan-96/64/orig 2025-09-07T07:28:18.6178785Z * [new branch] gh/ZhiweiYan-96/65/base -> origin/gh/ZhiweiYan-96/65/base 2025-09-07T07:28:18.6179611Z * [new branch] gh/ZhiweiYan-96/65/head -> origin/gh/ZhiweiYan-96/65/head 2025-09-07T07:28:18.6180452Z * [new branch] gh/ZhiweiYan-96/65/orig -> origin/gh/ZhiweiYan-96/65/orig 2025-09-07T07:28:18.6181678Z * [new branch] gh/ZhiweiYan-96/66/base -> origin/gh/ZhiweiYan-96/66/base 2025-09-07T07:28:18.6182498Z * [new branch] gh/ZhiweiYan-96/66/head -> origin/gh/ZhiweiYan-96/66/head 2025-09-07T07:28:18.6183702Z * [new branch] gh/ZhiweiYan-96/67/base -> origin/gh/ZhiweiYan-96/67/base 2025-09-07T07:28:18.6184456Z * [new branch] gh/ZhiweiYan-96/67/head -> origin/gh/ZhiweiYan-96/67/head 2025-09-07T07:28:18.6185536Z * [new branch] gh/ZhiweiYan-96/68/base -> origin/gh/ZhiweiYan-96/68/base 2025-09-07T07:28:18.6186414Z * [new branch] gh/ZhiweiYan-96/68/head -> origin/gh/ZhiweiYan-96/68/head 2025-09-07T07:28:18.6187294Z * [new branch] gh/ZhiweiYan-96/68/orig -> origin/gh/ZhiweiYan-96/68/orig 2025-09-07T07:28:18.6188800Z * [new branch] gh/aakhundov/1/base -> origin/gh/aakhundov/1/base 2025-09-07T07:28:18.6189735Z * [new branch] gh/aakhundov/1/head -> origin/gh/aakhundov/1/head 2025-09-07T07:28:18.6190763Z * [new branch] gh/aakhundov/2/base -> origin/gh/aakhundov/2/base 2025-09-07T07:28:18.6191670Z * [new branch] gh/aakhundov/2/head -> origin/gh/aakhundov/2/head 2025-09-07T07:28:18.6193094Z * [new branch] gh/aditew01/openblas -> origin/gh/aditew01/openblas 2025-09-07T07:28:18.6193847Z * [new branch] gh/aditew01/sbgemm -> origin/gh/aditew01/sbgemm 2025-09-07T07:28:18.6194722Z * [new branch] gh/aditew01/vecbf16 -> origin/gh/aditew01/vecbf16 2025-09-07T07:28:18.6196289Z * [new branch] gh/alexbrauckmann/paddedtensor_faketensor_init -> origin/gh/alexbrauckmann/paddedtensor_faketensor_init 2025-09-07T07:28:18.6197514Z * [new branch] gh/alexsamardzic/9/base -> origin/gh/alexsamardzic/9/base 2025-09-07T07:28:18.6198406Z * [new branch] gh/alexsamardzic/9/head -> origin/gh/alexsamardzic/9/head 2025-09-07T07:28:18.6199299Z * [new branch] gh/alexsamardzic/9/orig -> origin/gh/alexsamardzic/9/orig 2025-09-07T07:28:18.6200742Z * [new branch] gh/amjames/18/base -> origin/gh/amjames/18/base 2025-09-07T07:28:18.6201601Z * [new branch] gh/amjames/18/head -> origin/gh/amjames/18/head 2025-09-07T07:28:18.6202454Z * [new branch] gh/amjames/18/orig -> origin/gh/amjames/18/orig 2025-09-07T07:28:18.6204279Z * [new branch] gh/andrewor14/35/base -> origin/gh/andrewor14/35/base 2025-09-07T07:28:18.6205181Z * [new branch] gh/andrewor14/35/head -> origin/gh/andrewor14/35/head 2025-09-07T07:28:18.6205998Z * [new branch] gh/andrewor14/35/orig -> origin/gh/andrewor14/35/orig 2025-09-07T07:28:18.6207376Z * [new branch] gh/andrewor14/50/base -> origin/gh/andrewor14/50/base 2025-09-07T07:28:18.6208348Z * [new branch] gh/andrewor14/50/head -> origin/gh/andrewor14/50/head 2025-09-07T07:28:18.6209249Z * [new branch] gh/andrewor14/50/orig -> origin/gh/andrewor14/50/orig 2025-09-07T07:28:18.6210449Z * [new branch] gh/andrewor14/51/base -> origin/gh/andrewor14/51/base 2025-09-07T07:28:18.6211369Z * [new branch] gh/andrewor14/51/orig -> origin/gh/andrewor14/51/orig 2025-09-07T07:28:18.6213312Z * [new branch] gh/andyanwang/1/base -> origin/gh/andyanwang/1/base 2025-09-07T07:28:18.6213957Z * [new branch] gh/andyanwang/1/head -> origin/gh/andyanwang/1/head 2025-09-07T07:28:18.6214736Z * [new branch] gh/andyanwang/1/orig -> origin/gh/andyanwang/1/orig 2025-09-07T07:28:18.6216173Z * [new branch] gh/andyanwang/13/base -> origin/gh/andyanwang/13/base 2025-09-07T07:28:18.6217019Z * [new branch] gh/andyanwang/13/head -> origin/gh/andyanwang/13/head 2025-09-07T07:28:18.6218387Z * [new branch] gh/andyanwang/13/orig -> origin/gh/andyanwang/13/orig 2025-09-07T07:28:18.6219590Z * [new branch] gh/andyanwang/2/base -> origin/gh/andyanwang/2/base 2025-09-07T07:28:18.6220592Z * [new branch] gh/andyanwang/2/head -> origin/gh/andyanwang/2/head 2025-09-07T07:28:18.6221458Z * [new branch] gh/andyanwang/2/orig -> origin/gh/andyanwang/2/orig 2025-09-07T07:28:18.6222729Z * [new branch] gh/andyanwang/28/base -> origin/gh/andyanwang/28/base 2025-09-07T07:28:18.6223606Z * [new branch] gh/andyanwang/28/head -> origin/gh/andyanwang/28/head 2025-09-07T07:28:18.6224449Z * [new branch] gh/andyanwang/28/orig -> origin/gh/andyanwang/28/orig 2025-09-07T07:28:18.6225500Z * [new branch] gh/andyanwang/3/base -> origin/gh/andyanwang/3/base 2025-09-07T07:28:18.6226404Z * [new branch] gh/andyanwang/3/head -> origin/gh/andyanwang/3/head 2025-09-07T07:28:18.6227351Z * [new branch] gh/andyanwang/3/orig -> origin/gh/andyanwang/3/orig 2025-09-07T07:28:18.6228636Z * [new branch] gh/andyanwang/30/base -> origin/gh/andyanwang/30/base 2025-09-07T07:28:18.6229754Z * [new branch] gh/andyanwang/30/orig -> origin/gh/andyanwang/30/orig 2025-09-07T07:28:18.6230890Z * [new branch] gh/andyanwang/31/base -> origin/gh/andyanwang/31/base 2025-09-07T07:28:18.6231935Z * [new branch] gh/andyanwang/31/orig -> origin/gh/andyanwang/31/orig 2025-09-07T07:28:18.6233490Z * [new branch] gh/andyanwang/32/base -> origin/gh/andyanwang/32/base 2025-09-07T07:28:18.6234438Z * [new branch] gh/andyanwang/32/head -> origin/gh/andyanwang/32/head 2025-09-07T07:28:18.6235447Z * [new branch] gh/andyanwang/32/orig -> origin/gh/andyanwang/32/orig 2025-09-07T07:28:18.6236644Z * [new branch] gh/andyanwang/39/base -> origin/gh/andyanwang/39/base 2025-09-07T07:28:18.6237606Z * [new branch] gh/andyanwang/39/head -> origin/gh/andyanwang/39/head 2025-09-07T07:28:18.6238459Z * [new branch] gh/andyanwang/39/orig -> origin/gh/andyanwang/39/orig 2025-09-07T07:28:18.6239758Z * [new branch] gh/andyanwang/4/base -> origin/gh/andyanwang/4/base 2025-09-07T07:28:18.6240569Z * [new branch] gh/andyanwang/4/head -> origin/gh/andyanwang/4/head 2025-09-07T07:28:18.6241649Z * [new branch] gh/andyanwang/4/orig -> origin/gh/andyanwang/4/orig 2025-09-07T07:28:18.6243095Z * [new branch] gh/angelayi/107/base -> origin/gh/angelayi/107/base 2025-09-07T07:28:18.6244021Z * [new branch] gh/angelayi/107/head -> origin/gh/angelayi/107/head 2025-09-07T07:28:18.6245170Z * [new branch] gh/angelayi/111/base -> origin/gh/angelayi/111/base 2025-09-07T07:28:18.6246066Z * [new branch] gh/angelayi/111/head -> origin/gh/angelayi/111/head 2025-09-07T07:28:18.6246887Z * [new branch] gh/angelayi/111/orig -> origin/gh/angelayi/111/orig 2025-09-07T07:28:18.6248080Z * [new branch] gh/angelayi/112/base -> origin/gh/angelayi/112/base 2025-09-07T07:28:18.6248940Z * [new branch] gh/angelayi/112/head -> origin/gh/angelayi/112/head 2025-09-07T07:28:18.6249833Z * [new branch] gh/angelayi/112/orig -> origin/gh/angelayi/112/orig 2025-09-07T07:28:18.6251187Z * [new branch] gh/angelayi/113/base -> origin/gh/angelayi/113/base 2025-09-07T07:28:18.6251997Z * [new branch] gh/angelayi/113/head -> origin/gh/angelayi/113/head 2025-09-07T07:28:18.6252966Z * [new branch] gh/angelayi/113/orig -> origin/gh/angelayi/113/orig 2025-09-07T07:28:18.6254174Z * [new branch] gh/angelayi/114/base -> origin/gh/angelayi/114/base 2025-09-07T07:28:18.6254988Z * [new branch] gh/angelayi/114/head -> origin/gh/angelayi/114/head 2025-09-07T07:28:18.6255782Z * [new branch] gh/angelayi/114/orig -> origin/gh/angelayi/114/orig 2025-09-07T07:28:18.6257008Z * [new branch] gh/angelayi/115/base -> origin/gh/angelayi/115/base 2025-09-07T07:28:18.6257925Z * [new branch] gh/angelayi/115/head -> origin/gh/angelayi/115/head 2025-09-07T07:28:18.6258750Z * [new branch] gh/angelayi/115/orig -> origin/gh/angelayi/115/orig 2025-09-07T07:28:18.6260355Z * [new branch] gh/anijain2305/753/base -> origin/gh/anijain2305/753/base 2025-09-07T07:28:18.6261224Z * [new branch] gh/anijain2305/753/head -> origin/gh/anijain2305/753/head 2025-09-07T07:28:18.6262219Z * [new branch] gh/anijain2305/753/orig -> origin/gh/anijain2305/753/orig 2025-09-07T07:28:18.6263502Z * [new branch] gh/anijain2305/766/base -> origin/gh/anijain2305/766/base 2025-09-07T07:28:18.6264291Z * [new branch] gh/anijain2305/766/head -> origin/gh/anijain2305/766/head 2025-09-07T07:28:18.6265118Z * [new branch] gh/anijain2305/766/orig -> origin/gh/anijain2305/766/orig 2025-09-07T07:28:18.6267773Z * [new branch] gh/anijain2305/790/base -> origin/gh/anijain2305/790/base 2025-09-07T07:28:18.6268550Z * [new branch] gh/anijain2305/790/head -> origin/gh/anijain2305/790/head 2025-09-07T07:28:18.6269363Z * [new branch] gh/anijain2305/790/orig -> origin/gh/anijain2305/790/orig 2025-09-07T07:28:18.6270599Z * [new branch] gh/anijain2305/792/base -> origin/gh/anijain2305/792/base 2025-09-07T07:28:18.6271484Z * [new branch] gh/anijain2305/792/head -> origin/gh/anijain2305/792/head 2025-09-07T07:28:18.6272460Z * [new branch] gh/anijain2305/792/orig -> origin/gh/anijain2305/792/orig 2025-09-07T07:28:18.6273491Z * [new branch] gh/anijain2305/803/base -> origin/gh/anijain2305/803/base 2025-09-07T07:28:18.6274366Z * [new branch] gh/anijain2305/803/head -> origin/gh/anijain2305/803/head 2025-09-07T07:28:18.6275240Z * [new branch] gh/anijain2305/803/orig -> origin/gh/anijain2305/803/orig 2025-09-07T07:28:18.6276404Z * [new branch] gh/anijain2305/804/base -> origin/gh/anijain2305/804/base 2025-09-07T07:28:18.6277383Z * [new branch] gh/anijain2305/804/head -> origin/gh/anijain2305/804/head 2025-09-07T07:28:18.6278205Z * [new branch] gh/anijain2305/804/orig -> origin/gh/anijain2305/804/orig 2025-09-07T07:28:18.6279324Z * [new branch] gh/anijain2305/805/base -> origin/gh/anijain2305/805/base 2025-09-07T07:28:18.6280162Z * [new branch] gh/anijain2305/805/head -> origin/gh/anijain2305/805/head 2025-09-07T07:28:18.6281052Z * [new branch] gh/anijain2305/805/orig -> origin/gh/anijain2305/805/orig 2025-09-07T07:28:18.6282337Z * [new branch] gh/anijain2305/810/base -> origin/gh/anijain2305/810/base 2025-09-07T07:28:18.6283232Z * [new branch] gh/anijain2305/810/head -> origin/gh/anijain2305/810/head 2025-09-07T07:28:18.6284094Z * [new branch] gh/anijain2305/810/orig -> origin/gh/anijain2305/810/orig 2025-09-07T07:28:18.6285407Z * [new branch] gh/anijain2305/812/base -> origin/gh/anijain2305/812/base 2025-09-07T07:28:18.6286306Z * [new branch] gh/anijain2305/812/head -> origin/gh/anijain2305/812/head 2025-09-07T07:28:18.6287113Z * [new branch] gh/anijain2305/812/orig -> origin/gh/anijain2305/812/orig 2025-09-07T07:28:18.6288348Z * [new branch] gh/anijain2305/838/base -> origin/gh/anijain2305/838/base 2025-09-07T07:28:18.6289277Z * [new branch] gh/anijain2305/838/head -> origin/gh/anijain2305/838/head 2025-09-07T07:28:18.6290155Z * [new branch] gh/anijain2305/838/orig -> origin/gh/anijain2305/838/orig 2025-09-07T07:28:18.6291325Z * [new branch] gh/anijain2305/839/base -> origin/gh/anijain2305/839/base 2025-09-07T07:28:18.6292373Z * [new branch] gh/anijain2305/839/head -> origin/gh/anijain2305/839/head 2025-09-07T07:28:18.6293038Z * [new branch] gh/anijain2305/839/orig -> origin/gh/anijain2305/839/orig 2025-09-07T07:28:18.6294166Z * [new branch] gh/anijain2305/843/base -> origin/gh/anijain2305/843/base 2025-09-07T07:28:18.6295055Z * [new branch] gh/anijain2305/843/head -> origin/gh/anijain2305/843/head 2025-09-07T07:28:18.6295943Z * [new branch] gh/anijain2305/843/orig -> origin/gh/anijain2305/843/orig 2025-09-07T07:28:18.6297164Z * [new branch] gh/anijain2305/844/base -> origin/gh/anijain2305/844/base 2025-09-07T07:28:18.6298009Z * [new branch] gh/anijain2305/844/head -> origin/gh/anijain2305/844/head 2025-09-07T07:28:18.6298932Z * [new branch] gh/anijain2305/844/orig -> origin/gh/anijain2305/844/orig 2025-09-07T07:28:18.6300146Z * [new branch] gh/anijain2305/846/base -> origin/gh/anijain2305/846/base 2025-09-07T07:28:18.6301156Z * [new branch] gh/anijain2305/846/head -> origin/gh/anijain2305/846/head 2025-09-07T07:28:18.6301839Z * [new branch] gh/anijain2305/846/orig -> origin/gh/anijain2305/846/orig 2025-09-07T07:28:18.6303093Z * [new branch] gh/anijain2305/848/base -> origin/gh/anijain2305/848/base 2025-09-07T07:28:18.6304059Z * [new branch] gh/anijain2305/848/head -> origin/gh/anijain2305/848/head 2025-09-07T07:28:18.6304872Z * [new branch] gh/anijain2305/848/orig -> origin/gh/anijain2305/848/orig 2025-09-07T07:28:18.6306095Z * [new branch] gh/anijain2305/849/base -> origin/gh/anijain2305/849/base 2025-09-07T07:28:18.6306903Z * [new branch] gh/anijain2305/849/head -> origin/gh/anijain2305/849/head 2025-09-07T07:28:18.6307707Z * [new branch] gh/anijain2305/849/orig -> origin/gh/anijain2305/849/orig 2025-09-07T07:28:18.6309258Z * [new branch] gh/anijain2305/850/base -> origin/gh/anijain2305/850/base 2025-09-07T07:28:18.6310077Z * [new branch] gh/anijain2305/850/head -> origin/gh/anijain2305/850/head 2025-09-07T07:28:18.6310975Z * [new branch] gh/anijain2305/850/orig -> origin/gh/anijain2305/850/orig 2025-09-07T07:28:18.6312260Z * [new branch] gh/anijain2305/851/base -> origin/gh/anijain2305/851/base 2025-09-07T07:28:18.6313163Z * [new branch] gh/anijain2305/851/head -> origin/gh/anijain2305/851/head 2025-09-07T07:28:18.6313994Z * [new branch] gh/anijain2305/851/orig -> origin/gh/anijain2305/851/orig 2025-09-07T07:28:18.6315324Z * [new branch] gh/anijain2305/852/base -> origin/gh/anijain2305/852/base 2025-09-07T07:28:18.6316275Z * [new branch] gh/anijain2305/852/head -> origin/gh/anijain2305/852/head 2025-09-07T07:28:18.6317111Z * [new branch] gh/anijain2305/852/orig -> origin/gh/anijain2305/852/orig 2025-09-07T07:28:18.6318321Z * [new branch] gh/anijain2305/853/base -> origin/gh/anijain2305/853/base 2025-09-07T07:28:18.6319143Z * [new branch] gh/anijain2305/853/head -> origin/gh/anijain2305/853/head 2025-09-07T07:28:18.6319991Z * [new branch] gh/anijain2305/853/orig -> origin/gh/anijain2305/853/orig 2025-09-07T07:28:18.6321217Z * [new branch] gh/anijain2305/854/base -> origin/gh/anijain2305/854/base 2025-09-07T07:28:18.6322113Z * [new branch] gh/anijain2305/854/head -> origin/gh/anijain2305/854/head 2025-09-07T07:28:18.6323294Z * [new branch] gh/anijain2305/854/orig -> origin/gh/anijain2305/854/orig 2025-09-07T07:28:18.6324299Z * [new branch] gh/anijain2305/855/base -> origin/gh/anijain2305/855/base 2025-09-07T07:28:18.6325257Z * [new branch] gh/anijain2305/855/head -> origin/gh/anijain2305/855/head 2025-09-07T07:28:18.6326082Z * [new branch] gh/anijain2305/855/orig -> origin/gh/anijain2305/855/orig 2025-09-07T07:28:18.6327345Z * [new branch] gh/anijain2305/856/base -> origin/gh/anijain2305/856/base 2025-09-07T07:28:18.6328186Z * [new branch] gh/anijain2305/856/head -> origin/gh/anijain2305/856/head 2025-09-07T07:28:18.6329049Z * [new branch] gh/anijain2305/856/orig -> origin/gh/anijain2305/856/orig 2025-09-07T07:28:18.6330188Z * [new branch] gh/anijain2305/857/base -> origin/gh/anijain2305/857/base 2025-09-07T07:28:18.6331113Z * [new branch] gh/anijain2305/857/head -> origin/gh/anijain2305/857/head 2025-09-07T07:28:18.6331961Z * [new branch] gh/anijain2305/857/orig -> origin/gh/anijain2305/857/orig 2025-09-07T07:28:18.6333168Z * [new branch] gh/anijain2305/858/base -> origin/gh/anijain2305/858/base 2025-09-07T07:28:18.6334012Z * [new branch] gh/anijain2305/858/head -> origin/gh/anijain2305/858/head 2025-09-07T07:28:18.6335017Z * [new branch] gh/anijain2305/858/orig -> origin/gh/anijain2305/858/orig 2025-09-07T07:28:18.6336129Z * [new branch] gh/anijain2305/859/base -> origin/gh/anijain2305/859/base 2025-09-07T07:28:18.6337030Z * [new branch] gh/anijain2305/859/head -> origin/gh/anijain2305/859/head 2025-09-07T07:28:18.6337883Z * [new branch] gh/anijain2305/859/orig -> origin/gh/anijain2305/859/orig 2025-09-07T07:28:18.6339045Z * [new branch] gh/anijain2305/860/base -> origin/gh/anijain2305/860/base 2025-09-07T07:28:18.6339951Z * [new branch] gh/anijain2305/860/head -> origin/gh/anijain2305/860/head 2025-09-07T07:28:18.6340791Z * [new branch] gh/anijain2305/860/orig -> origin/gh/anijain2305/860/orig 2025-09-07T07:28:18.6342031Z * [new branch] gh/anijain2305/861/base -> origin/gh/anijain2305/861/base 2025-09-07T07:28:18.6342845Z * [new branch] gh/anijain2305/861/head -> origin/gh/anijain2305/861/head 2025-09-07T07:28:18.6343742Z * [new branch] gh/anijain2305/861/orig -> origin/gh/anijain2305/861/orig 2025-09-07T07:28:18.6344982Z * [new branch] gh/anijain2305/862/base -> origin/gh/anijain2305/862/base 2025-09-07T07:28:18.6345865Z * [new branch] gh/anijain2305/862/head -> origin/gh/anijain2305/862/head 2025-09-07T07:28:18.6346785Z * [new branch] gh/anijain2305/862/orig -> origin/gh/anijain2305/862/orig 2025-09-07T07:28:18.6348031Z * [new branch] gh/anijain2305/863/base -> origin/gh/anijain2305/863/base 2025-09-07T07:28:18.6349072Z * [new branch] gh/anijain2305/863/head -> origin/gh/anijain2305/863/head 2025-09-07T07:28:18.6349977Z * [new branch] gh/anijain2305/863/orig -> origin/gh/anijain2305/863/orig 2025-09-07T07:28:18.6351279Z * [new branch] gh/anijain2305/864/base -> origin/gh/anijain2305/864/base 2025-09-07T07:28:18.6352146Z * [new branch] gh/anijain2305/864/head -> origin/gh/anijain2305/864/head 2025-09-07T07:28:18.6352941Z * [new branch] gh/anijain2305/864/orig -> origin/gh/anijain2305/864/orig 2025-09-07T07:28:18.6354210Z * [new branch] gh/anijain2305/865/base -> origin/gh/anijain2305/865/base 2025-09-07T07:28:18.6355064Z * [new branch] gh/anijain2305/865/head -> origin/gh/anijain2305/865/head 2025-09-07T07:28:18.6355932Z * [new branch] gh/anijain2305/865/orig -> origin/gh/anijain2305/865/orig 2025-09-07T07:28:18.6357146Z * [new branch] gh/anijain2305/866/base -> origin/gh/anijain2305/866/base 2025-09-07T07:28:18.6358026Z * [new branch] gh/anijain2305/866/head -> origin/gh/anijain2305/866/head 2025-09-07T07:28:18.6358895Z * [new branch] gh/anijain2305/866/orig -> origin/gh/anijain2305/866/orig 2025-09-07T07:28:18.6360549Z * [new branch] gh/anjali411/216/base -> origin/gh/anjali411/216/base 2025-09-07T07:28:18.6361432Z * [new branch] gh/anjali411/216/head -> origin/gh/anjali411/216/head 2025-09-07T07:28:18.6362293Z * [new branch] gh/anjali411/216/orig -> origin/gh/anjali411/216/orig 2025-09-07T07:28:18.6363781Z * [new branch] gh/ankitageorge/13/base -> origin/gh/ankitageorge/13/base 2025-09-07T07:28:18.6364643Z * [new branch] gh/ankitageorge/13/head -> origin/gh/ankitageorge/13/head 2025-09-07T07:28:18.6365666Z * [new branch] gh/ankitageorge/13/orig -> origin/gh/ankitageorge/13/orig 2025-09-07T07:28:18.6366975Z * [new branch] gh/ankitageorge/14/base -> origin/gh/ankitageorge/14/base 2025-09-07T07:28:18.6367782Z * [new branch] gh/ankitageorge/14/head -> origin/gh/ankitageorge/14/head 2025-09-07T07:28:18.6368848Z * [new branch] gh/ankitageorge/14/orig -> origin/gh/ankitageorge/14/orig 2025-09-07T07:28:18.6370282Z * [new branch] gh/ankitageorge/15/base -> origin/gh/ankitageorge/15/base 2025-09-07T07:28:18.6370986Z * [new branch] gh/ankitageorge/15/head -> origin/gh/ankitageorge/15/head 2025-09-07T07:28:18.6371897Z * [new branch] gh/ankitageorge/15/orig -> origin/gh/ankitageorge/15/orig 2025-09-07T07:28:18.6373508Z * [new branch] gh/ankitageorge/16/base -> origin/gh/ankitageorge/16/base 2025-09-07T07:28:18.6374413Z * [new branch] gh/ankitageorge/16/head -> origin/gh/ankitageorge/16/head 2025-09-07T07:28:18.6375265Z * [new branch] gh/ankitageorge/16/orig -> origin/gh/ankitageorge/16/orig 2025-09-07T07:28:18.6376688Z * [new branch] gh/ankitageorge/17/base -> origin/gh/ankitageorge/17/base 2025-09-07T07:28:18.6377472Z * [new branch] gh/ankitageorge/17/head -> origin/gh/ankitageorge/17/head 2025-09-07T07:28:18.6378353Z * [new branch] gh/ankitageorge/17/orig -> origin/gh/ankitageorge/17/orig 2025-09-07T07:28:18.6379768Z * [new branch] gh/ankitageorge/21/base -> origin/gh/ankitageorge/21/base 2025-09-07T07:28:18.6380549Z * [new branch] gh/ankitageorge/21/head -> origin/gh/ankitageorge/21/head 2025-09-07T07:28:18.6381456Z * [new branch] gh/ankitageorge/21/orig -> origin/gh/ankitageorge/21/orig 2025-09-07T07:28:18.6383057Z * [new branch] gh/anshul-si/1/base -> origin/gh/anshul-si/1/base 2025-09-07T07:28:18.6383940Z * [new branch] gh/anshul-si/1/head -> origin/gh/anshul-si/1/head 2025-09-07T07:28:18.6385212Z * [new branch] gh/anshul-si/15/base -> origin/gh/anshul-si/15/base 2025-09-07T07:28:18.6386110Z * [new branch] gh/anshul-si/15/head -> origin/gh/anshul-si/15/head 2025-09-07T07:28:18.6386992Z * [new branch] gh/anshul-si/15/orig -> origin/gh/anshul-si/15/orig 2025-09-07T07:28:18.6388334Z * [new branch] gh/anshul-si/16/base -> origin/gh/anshul-si/16/base 2025-09-07T07:28:18.6389157Z * [new branch] gh/anshul-si/16/head -> origin/gh/anshul-si/16/head 2025-09-07T07:28:18.6390058Z * [new branch] gh/anshul-si/16/orig -> origin/gh/anshul-si/16/orig 2025-09-07T07:28:18.6391309Z * [new branch] gh/anshul-si/17/base -> origin/gh/anshul-si/17/base 2025-09-07T07:28:18.6392302Z * [new branch] gh/anshul-si/17/head -> origin/gh/anshul-si/17/head 2025-09-07T07:28:18.6393309Z * [new branch] gh/anshul-si/17/orig -> origin/gh/anshul-si/17/orig 2025-09-07T07:28:18.6394571Z * [new branch] gh/anshul-si/18/base -> origin/gh/anshul-si/18/base 2025-09-07T07:28:18.6395542Z * [new branch] gh/anshul-si/18/head -> origin/gh/anshul-si/18/head 2025-09-07T07:28:18.6396471Z * [new branch] gh/anshul-si/18/orig -> origin/gh/anshul-si/18/orig 2025-09-07T07:28:18.6397778Z * [new branch] gh/anshul-si/19/base -> origin/gh/anshul-si/19/base 2025-09-07T07:28:18.6398957Z * [new branch] gh/anshul-si/19/head -> origin/gh/anshul-si/19/head 2025-09-07T07:28:18.6399567Z * [new branch] gh/anshul-si/19/orig -> origin/gh/anshul-si/19/orig 2025-09-07T07:28:18.6400830Z * [new branch] gh/anshul-si/2/base -> origin/gh/anshul-si/2/base 2025-09-07T07:28:18.6401701Z * [new branch] gh/anshul-si/2/head -> origin/gh/anshul-si/2/head 2025-09-07T07:28:18.6403062Z * [new branch] gh/anshul-si/20/base -> origin/gh/anshul-si/20/base 2025-09-07T07:28:18.6403934Z * [new branch] gh/anshul-si/20/head -> origin/gh/anshul-si/20/head 2025-09-07T07:28:18.6405013Z * [new branch] gh/anshul-si/20/orig -> origin/gh/anshul-si/20/orig 2025-09-07T07:28:18.6405929Z * [new branch] gh/anshul-si/21/base -> origin/gh/anshul-si/21/base 2025-09-07T07:28:18.6406901Z * [new branch] gh/anshul-si/21/head -> origin/gh/anshul-si/21/head 2025-09-07T07:28:18.6407673Z * [new branch] gh/anshul-si/21/orig -> origin/gh/anshul-si/21/orig 2025-09-07T07:28:18.6408948Z * [new branch] gh/anshul-si/22/base -> origin/gh/anshul-si/22/base 2025-09-07T07:28:18.6409895Z * [new branch] gh/anshul-si/22/head -> origin/gh/anshul-si/22/head 2025-09-07T07:28:18.6410739Z * [new branch] gh/anshul-si/22/orig -> origin/gh/anshul-si/22/orig 2025-09-07T07:28:18.6411981Z * [new branch] gh/anshul-si/23/base -> origin/gh/anshul-si/23/base 2025-09-07T07:28:18.6412880Z * [new branch] gh/anshul-si/23/head -> origin/gh/anshul-si/23/head 2025-09-07T07:28:18.6413763Z * [new branch] gh/anshul-si/23/orig -> origin/gh/anshul-si/23/orig 2025-09-07T07:28:18.6414947Z * [new branch] gh/anshul-si/24/base -> origin/gh/anshul-si/24/base 2025-09-07T07:28:18.6415899Z * [new branch] gh/anshul-si/24/head -> origin/gh/anshul-si/24/head 2025-09-07T07:28:18.6416784Z * [new branch] gh/anshul-si/24/orig -> origin/gh/anshul-si/24/orig 2025-09-07T07:28:18.6417976Z * [new branch] gh/anshul-si/25/base -> origin/gh/anshul-si/25/base 2025-09-07T07:28:18.6418879Z * [new branch] gh/anshul-si/25/head -> origin/gh/anshul-si/25/head 2025-09-07T07:28:18.6419744Z * [new branch] gh/anshul-si/25/orig -> origin/gh/anshul-si/25/orig 2025-09-07T07:28:18.6420952Z * [new branch] gh/anshul-si/26/base -> origin/gh/anshul-si/26/base 2025-09-07T07:28:18.6421816Z * [new branch] gh/anshul-si/26/head -> origin/gh/anshul-si/26/head 2025-09-07T07:28:18.6422697Z * [new branch] gh/anshul-si/26/orig -> origin/gh/anshul-si/26/orig 2025-09-07T07:28:18.6423923Z * [new branch] gh/anshul-si/27/base -> origin/gh/anshul-si/27/base 2025-09-07T07:28:18.6424788Z * [new branch] gh/anshul-si/27/head -> origin/gh/anshul-si/27/head 2025-09-07T07:28:18.6425641Z * [new branch] gh/anshul-si/27/orig -> origin/gh/anshul-si/27/orig 2025-09-07T07:28:18.6426774Z * [new branch] gh/anshul-si/28/base -> origin/gh/anshul-si/28/base 2025-09-07T07:28:18.6427642Z * [new branch] gh/anshul-si/28/head -> origin/gh/anshul-si/28/head 2025-09-07T07:28:18.6428489Z * [new branch] gh/anshul-si/28/orig -> origin/gh/anshul-si/28/orig 2025-09-07T07:28:18.6430144Z * [new branch] gh/anshul-si/29/base -> origin/gh/anshul-si/29/base 2025-09-07T07:28:18.6431216Z * [new branch] gh/anshul-si/29/head -> origin/gh/anshul-si/29/head 2025-09-07T07:28:18.6432106Z * [new branch] gh/anshul-si/29/orig -> origin/gh/anshul-si/29/orig 2025-09-07T07:28:18.6433212Z * [new branch] gh/anshul-si/3/base -> origin/gh/anshul-si/3/base 2025-09-07T07:28:18.6434031Z * [new branch] gh/anshul-si/3/head -> origin/gh/anshul-si/3/head 2025-09-07T07:28:18.6435126Z * [new branch] gh/anshul-si/4/base -> origin/gh/anshul-si/4/base 2025-09-07T07:28:18.6435921Z * [new branch] gh/anshul-si/4/head -> origin/gh/anshul-si/4/head 2025-09-07T07:28:18.6437317Z * [new branch] gh/anshul-si/5/base -> origin/gh/anshul-si/5/base 2025-09-07T07:28:18.6438248Z * [new branch] gh/anshul-si/5/head -> origin/gh/anshul-si/5/head 2025-09-07T07:28:18.6440007Z * [new branch] gh/aorenste/132/base -> origin/gh/aorenste/132/base 2025-09-07T07:28:18.6440812Z * [new branch] gh/aorenste/132/head -> origin/gh/aorenste/132/head 2025-09-07T07:28:18.6442346Z * [new branch] gh/bdhirsh/650/base -> origin/gh/bdhirsh/650/base 2025-09-07T07:28:18.6443537Z * [new branch] gh/bdhirsh/650/head -> origin/gh/bdhirsh/650/head 2025-09-07T07:28:18.6444517Z * [new branch] gh/bdhirsh/650/orig -> origin/gh/bdhirsh/650/orig 2025-09-07T07:28:18.6445793Z * [new branch] gh/bdhirsh/663/base -> origin/gh/bdhirsh/663/base 2025-09-07T07:28:18.6446662Z * [new branch] gh/bdhirsh/663/head -> origin/gh/bdhirsh/663/head 2025-09-07T07:28:18.6447539Z * [new branch] gh/bdhirsh/663/orig -> origin/gh/bdhirsh/663/orig 2025-09-07T07:28:18.6448929Z * [new branch] gh/bdhirsh/665/base -> origin/gh/bdhirsh/665/base 2025-09-07T07:28:18.6449744Z * [new branch] gh/bdhirsh/665/head -> origin/gh/bdhirsh/665/head 2025-09-07T07:28:18.6450595Z * [new branch] gh/bdhirsh/665/orig -> origin/gh/bdhirsh/665/orig 2025-09-07T07:28:18.6452096Z * [new branch] gh/bdhirsh/666/base -> origin/gh/bdhirsh/666/base 2025-09-07T07:28:18.6452981Z * [new branch] gh/bdhirsh/666/head -> origin/gh/bdhirsh/666/head 2025-09-07T07:28:18.6453887Z * [new branch] gh/bdhirsh/666/orig -> origin/gh/bdhirsh/666/orig 2025-09-07T07:28:18.6455296Z * [new branch] gh/bdhirsh/667/base -> origin/gh/bdhirsh/667/base 2025-09-07T07:28:18.6456129Z * [new branch] gh/bdhirsh/667/head -> origin/gh/bdhirsh/667/head 2025-09-07T07:28:18.6457081Z * [new branch] gh/bdhirsh/667/orig -> origin/gh/bdhirsh/667/orig 2025-09-07T07:28:18.6458233Z * [new branch] gh/bdhirsh/668/base -> origin/gh/bdhirsh/668/base 2025-09-07T07:28:18.6459106Z * [new branch] gh/bdhirsh/668/head -> origin/gh/bdhirsh/668/head 2025-09-07T07:28:18.6459971Z * [new branch] gh/bdhirsh/668/orig -> origin/gh/bdhirsh/668/orig 2025-09-07T07:28:18.6461318Z * [new branch] gh/bdhirsh/669/base -> origin/gh/bdhirsh/669/base 2025-09-07T07:28:18.6462132Z * [new branch] gh/bdhirsh/669/head -> origin/gh/bdhirsh/669/head 2025-09-07T07:28:18.6462959Z * [new branch] gh/bdhirsh/669/orig -> origin/gh/bdhirsh/669/orig 2025-09-07T07:28:18.6464289Z * [new branch] gh/bdhirsh/670/base -> origin/gh/bdhirsh/670/base 2025-09-07T07:28:18.6465254Z * [new branch] gh/bdhirsh/670/head -> origin/gh/bdhirsh/670/head 2025-09-07T07:28:18.6466342Z * [new branch] gh/bdhirsh/670/orig -> origin/gh/bdhirsh/670/orig 2025-09-07T07:28:18.6467898Z * [new branch] gh/benjaminglass1/100/base -> origin/gh/benjaminglass1/100/base 2025-09-07T07:28:18.6468853Z * [new branch] gh/benjaminglass1/100/head -> origin/gh/benjaminglass1/100/head 2025-09-07T07:28:18.6469806Z * [new branch] gh/benjaminglass1/100/orig -> origin/gh/benjaminglass1/100/orig 2025-09-07T07:28:18.6471100Z * [new branch] gh/benjaminglass1/101/base -> origin/gh/benjaminglass1/101/base 2025-09-07T07:28:18.6471979Z * [new branch] gh/benjaminglass1/101/head -> origin/gh/benjaminglass1/101/head 2025-09-07T07:28:18.6472894Z * [new branch] gh/benjaminglass1/101/orig -> origin/gh/benjaminglass1/101/orig 2025-09-07T07:28:18.6474135Z * [new branch] gh/benjaminglass1/102/base -> origin/gh/benjaminglass1/102/base 2025-09-07T07:28:18.6474978Z * [new branch] gh/benjaminglass1/102/head -> origin/gh/benjaminglass1/102/head 2025-09-07T07:28:18.6475883Z * [new branch] gh/benjaminglass1/102/orig -> origin/gh/benjaminglass1/102/orig 2025-09-07T07:28:18.6477108Z * [new branch] gh/benjaminglass1/103/base -> origin/gh/benjaminglass1/103/base 2025-09-07T07:28:18.6477931Z * [new branch] gh/benjaminglass1/103/head -> origin/gh/benjaminglass1/103/head 2025-09-07T07:28:18.6478773Z * [new branch] gh/benjaminglass1/103/orig -> origin/gh/benjaminglass1/103/orig 2025-09-07T07:28:18.6480233Z * [new branch] gh/benjaminglass1/104/base -> origin/gh/benjaminglass1/104/base 2025-09-07T07:28:18.6480910Z * [new branch] gh/benjaminglass1/104/head -> origin/gh/benjaminglass1/104/head 2025-09-07T07:28:18.6481795Z * [new branch] gh/benjaminglass1/104/orig -> origin/gh/benjaminglass1/104/orig 2025-09-07T07:28:18.6483017Z * [new branch] gh/benjaminglass1/105/base -> origin/gh/benjaminglass1/105/base 2025-09-07T07:28:18.6483829Z * [new branch] gh/benjaminglass1/105/head -> origin/gh/benjaminglass1/105/head 2025-09-07T07:28:18.6484709Z * [new branch] gh/benjaminglass1/105/orig -> origin/gh/benjaminglass1/105/orig 2025-09-07T07:28:18.6485855Z * [new branch] gh/benjaminglass1/106/base -> origin/gh/benjaminglass1/106/base 2025-09-07T07:28:18.6486684Z * [new branch] gh/benjaminglass1/106/head -> origin/gh/benjaminglass1/106/head 2025-09-07T07:28:18.6487673Z * [new branch] gh/benjaminglass1/106/orig -> origin/gh/benjaminglass1/106/orig 2025-09-07T07:28:18.6488816Z * [new branch] gh/benjaminglass1/79/base -> origin/gh/benjaminglass1/79/base 2025-09-07T07:28:18.6489706Z * [new branch] gh/benjaminglass1/79/head -> origin/gh/benjaminglass1/79/head 2025-09-07T07:28:18.6490562Z * [new branch] gh/benjaminglass1/79/orig -> origin/gh/benjaminglass1/79/orig 2025-09-07T07:28:18.6491716Z * [new branch] gh/benjaminglass1/86/base -> origin/gh/benjaminglass1/86/base 2025-09-07T07:28:18.6492573Z * [new branch] gh/benjaminglass1/86/head -> origin/gh/benjaminglass1/86/head 2025-09-07T07:28:18.6493506Z * [new branch] gh/benjaminglass1/86/orig -> origin/gh/benjaminglass1/86/orig 2025-09-07T07:28:18.6494715Z * [new branch] gh/benjaminglass1/89/base -> origin/gh/benjaminglass1/89/base 2025-09-07T07:28:18.6495560Z * [new branch] gh/benjaminglass1/89/head -> origin/gh/benjaminglass1/89/head 2025-09-07T07:28:18.6496415Z * [new branch] gh/benjaminglass1/89/orig -> origin/gh/benjaminglass1/89/orig 2025-09-07T07:28:18.6497548Z * [new branch] gh/benjaminglass1/91/base -> origin/gh/benjaminglass1/91/base 2025-09-07T07:28:18.6498469Z * [new branch] gh/benjaminglass1/91/head -> origin/gh/benjaminglass1/91/head 2025-09-07T07:28:18.6499250Z * [new branch] gh/benjaminglass1/91/orig -> origin/gh/benjaminglass1/91/orig 2025-09-07T07:28:18.6500499Z * [new branch] gh/benjaminglass1/93/base -> origin/gh/benjaminglass1/93/base 2025-09-07T07:28:18.6501364Z * [new branch] gh/benjaminglass1/93/head -> origin/gh/benjaminglass1/93/head 2025-09-07T07:28:18.6502233Z * [new branch] gh/benjaminglass1/93/orig -> origin/gh/benjaminglass1/93/orig 2025-09-07T07:28:18.6503388Z * [new branch] gh/benjaminglass1/95/base -> origin/gh/benjaminglass1/95/base 2025-09-07T07:28:18.6504231Z * [new branch] gh/benjaminglass1/95/head -> origin/gh/benjaminglass1/95/head 2025-09-07T07:28:18.6505111Z * [new branch] gh/benjaminglass1/95/orig -> origin/gh/benjaminglass1/95/orig 2025-09-07T07:28:18.6506321Z * [new branch] gh/benjaminglass1/97/base -> origin/gh/benjaminglass1/97/base 2025-09-07T07:28:18.6507223Z * [new branch] gh/benjaminglass1/97/head -> origin/gh/benjaminglass1/97/head 2025-09-07T07:28:18.6508171Z * [new branch] gh/benjaminglass1/97/orig -> origin/gh/benjaminglass1/97/orig 2025-09-07T07:28:18.6509372Z * [new branch] gh/benjaminglass1/99/base -> origin/gh/benjaminglass1/99/base 2025-09-07T07:28:18.6510183Z * [new branch] gh/benjaminglass1/99/head -> origin/gh/benjaminglass1/99/head 2025-09-07T07:28:18.6511129Z * [new branch] gh/benjaminglass1/99/orig -> origin/gh/benjaminglass1/99/orig 2025-09-07T07:28:18.6512713Z * [new branch] gh/bobrenjc93/514/base -> origin/gh/bobrenjc93/514/base 2025-09-07T07:28:18.6513479Z * [new branch] gh/bobrenjc93/514/head -> origin/gh/bobrenjc93/514/head 2025-09-07T07:28:18.6514311Z * [new branch] gh/bobrenjc93/514/orig -> origin/gh/bobrenjc93/514/orig 2025-09-07T07:28:18.6516471Z * [new branch] gh/bobrenjc93/521/base -> origin/gh/bobrenjc93/521/base 2025-09-07T07:28:18.6516882Z * [new branch] gh/bobrenjc93/521/head -> origin/gh/bobrenjc93/521/head 2025-09-07T07:28:18.6517369Z * [new branch] gh/bobrenjc93/521/orig -> origin/gh/bobrenjc93/521/orig 2025-09-07T07:28:18.6518558Z * [new branch] gh/bobrenjc93/522/base -> origin/gh/bobrenjc93/522/base 2025-09-07T07:28:18.6519345Z * [new branch] gh/bobrenjc93/522/head -> origin/gh/bobrenjc93/522/head 2025-09-07T07:28:18.6520222Z * [new branch] gh/bobrenjc93/522/orig -> origin/gh/bobrenjc93/522/orig 2025-09-07T07:28:18.6521386Z * [new branch] gh/bobrenjc93/525/base -> origin/gh/bobrenjc93/525/base 2025-09-07T07:28:18.6522247Z * [new branch] gh/bobrenjc93/525/head -> origin/gh/bobrenjc93/525/head 2025-09-07T07:28:18.6523104Z * [new branch] gh/bobrenjc93/525/orig -> origin/gh/bobrenjc93/525/orig 2025-09-07T07:28:18.6524458Z * [new branch] gh/bobrenjc93/526/base -> origin/gh/bobrenjc93/526/base 2025-09-07T07:28:18.6525328Z * [new branch] gh/bobrenjc93/526/head -> origin/gh/bobrenjc93/526/head 2025-09-07T07:28:18.6526161Z * [new branch] gh/bobrenjc93/526/orig -> origin/gh/bobrenjc93/526/orig 2025-09-07T07:28:18.6527395Z * [new branch] gh/bobrenjc93/527/base -> origin/gh/bobrenjc93/527/base 2025-09-07T07:28:18.6528486Z * [new branch] gh/bobrenjc93/527/head -> origin/gh/bobrenjc93/527/head 2025-09-07T07:28:18.6529387Z * [new branch] gh/bobrenjc93/527/orig -> origin/gh/bobrenjc93/527/orig 2025-09-07T07:28:18.6530542Z * [new branch] gh/bobrenjc93/528/base -> origin/gh/bobrenjc93/528/base 2025-09-07T07:28:18.6531432Z * [new branch] gh/bobrenjc93/528/head -> origin/gh/bobrenjc93/528/head 2025-09-07T07:28:18.6532278Z * [new branch] gh/bobrenjc93/528/orig -> origin/gh/bobrenjc93/528/orig 2025-09-07T07:28:18.6533461Z * [new branch] gh/bobrenjc93/529/base -> origin/gh/bobrenjc93/529/base 2025-09-07T07:28:18.6534269Z * [new branch] gh/bobrenjc93/529/head -> origin/gh/bobrenjc93/529/head 2025-09-07T07:28:18.6535169Z * [new branch] gh/bobrenjc93/529/orig -> origin/gh/bobrenjc93/529/orig 2025-09-07T07:28:18.6536689Z * [new branch] gh/bobrenjc93/535/base -> origin/gh/bobrenjc93/535/base 2025-09-07T07:28:18.6537153Z * [new branch] gh/bobrenjc93/535/head -> origin/gh/bobrenjc93/535/head 2025-09-07T07:28:18.6538034Z * [new branch] gh/bobrenjc93/535/orig -> origin/gh/bobrenjc93/535/orig 2025-09-07T07:28:18.6539297Z * [new branch] gh/bobrenjc93/537/base -> origin/gh/bobrenjc93/537/base 2025-09-07T07:28:18.6540329Z * [new branch] gh/bobrenjc93/537/head -> origin/gh/bobrenjc93/537/head 2025-09-07T07:28:18.6541173Z * [new branch] gh/bobrenjc93/537/orig -> origin/gh/bobrenjc93/537/orig 2025-09-07T07:28:18.6542566Z * [new branch] gh/bobrenjc93/539/base -> origin/gh/bobrenjc93/539/base 2025-09-07T07:28:18.6543399Z * [new branch] gh/bobrenjc93/539/head -> origin/gh/bobrenjc93/539/head 2025-09-07T07:28:18.6544348Z * [new branch] gh/bobrenjc93/539/orig -> origin/gh/bobrenjc93/539/orig 2025-09-07T07:28:18.6545573Z * [new branch] gh/bobrenjc93/540/base -> origin/gh/bobrenjc93/540/base 2025-09-07T07:28:18.6546481Z * [new branch] gh/bobrenjc93/540/head -> origin/gh/bobrenjc93/540/head 2025-09-07T07:28:18.6547446Z * [new branch] gh/bobrenjc93/540/orig -> origin/gh/bobrenjc93/540/orig 2025-09-07T07:28:18.6548582Z * [new branch] gh/bobrenjc93/541/base -> origin/gh/bobrenjc93/541/base 2025-09-07T07:28:18.6549482Z * [new branch] gh/bobrenjc93/541/head -> origin/gh/bobrenjc93/541/head 2025-09-07T07:28:18.6550338Z * [new branch] gh/bobrenjc93/541/orig -> origin/gh/bobrenjc93/541/orig 2025-09-07T07:28:18.6551411Z * [new branch] gh/bobrenjc93/542/base -> origin/gh/bobrenjc93/542/base 2025-09-07T07:28:18.6552300Z * [new branch] gh/bobrenjc93/542/head -> origin/gh/bobrenjc93/542/head 2025-09-07T07:28:18.6553235Z * [new branch] gh/bobrenjc93/542/orig -> origin/gh/bobrenjc93/542/orig 2025-09-07T07:28:18.6554405Z * [new branch] gh/bobrenjc93/543/base -> origin/gh/bobrenjc93/543/base 2025-09-07T07:28:18.6555211Z * [new branch] gh/bobrenjc93/543/head -> origin/gh/bobrenjc93/543/head 2025-09-07T07:28:18.6556137Z * [new branch] gh/bobrenjc93/543/orig -> origin/gh/bobrenjc93/543/orig 2025-09-07T07:28:18.6557199Z * [new branch] gh/bobrenjc93/544/base -> origin/gh/bobrenjc93/544/base 2025-09-07T07:28:18.6558047Z * [new branch] gh/bobrenjc93/544/head -> origin/gh/bobrenjc93/544/head 2025-09-07T07:28:18.6558916Z * [new branch] gh/bobrenjc93/544/orig -> origin/gh/bobrenjc93/544/orig 2025-09-07T07:28:18.6560310Z * [new branch] gh/bobrenjc93/545/base -> origin/gh/bobrenjc93/545/base 2025-09-07T07:28:18.6561290Z * [new branch] gh/bobrenjc93/545/head -> origin/gh/bobrenjc93/545/head 2025-09-07T07:28:18.6562193Z * [new branch] gh/bobrenjc93/545/orig -> origin/gh/bobrenjc93/545/orig 2025-09-07T07:28:18.6563498Z * [new branch] gh/bobrenjc93/546/base -> origin/gh/bobrenjc93/546/base 2025-09-07T07:28:18.6564447Z * [new branch] gh/bobrenjc93/546/head -> origin/gh/bobrenjc93/546/head 2025-09-07T07:28:18.6565349Z * [new branch] gh/bobrenjc93/546/orig -> origin/gh/bobrenjc93/546/orig 2025-09-07T07:28:18.6567260Z * [new branch] gh/bobrenjc93/547/base -> origin/gh/bobrenjc93/547/base 2025-09-07T07:28:18.6568209Z * [new branch] gh/bobrenjc93/547/head -> origin/gh/bobrenjc93/547/head 2025-09-07T07:28:18.6569151Z * [new branch] gh/bobrenjc93/547/orig -> origin/gh/bobrenjc93/547/orig 2025-09-07T07:28:18.6570262Z * [new branch] gh/bobrenjc93/548/base -> origin/gh/bobrenjc93/548/base 2025-09-07T07:28:18.6571127Z * [new branch] gh/bobrenjc93/548/head -> origin/gh/bobrenjc93/548/head 2025-09-07T07:28:18.6571978Z * [new branch] gh/bobrenjc93/548/orig -> origin/gh/bobrenjc93/548/orig 2025-09-07T07:28:18.6573097Z * [new branch] gh/bobrenjc93/549/base -> origin/gh/bobrenjc93/549/base 2025-09-07T07:28:18.6574002Z * [new branch] gh/bobrenjc93/549/head -> origin/gh/bobrenjc93/549/head 2025-09-07T07:28:18.6574948Z * [new branch] gh/bobrenjc93/549/orig -> origin/gh/bobrenjc93/549/orig 2025-09-07T07:28:18.6576344Z * [new branch] gh/bobrenjc93/550/base -> origin/gh/bobrenjc93/550/base 2025-09-07T07:28:18.6577265Z * [new branch] gh/bobrenjc93/550/head -> origin/gh/bobrenjc93/550/head 2025-09-07T07:28:18.6578224Z * [new branch] gh/bobrenjc93/550/orig -> origin/gh/bobrenjc93/550/orig 2025-09-07T07:28:18.6579658Z * [new branch] gh/bobrenjc93/551/base -> origin/gh/bobrenjc93/551/base 2025-09-07T07:28:18.6580575Z * [new branch] gh/bobrenjc93/551/head -> origin/gh/bobrenjc93/551/head 2025-09-07T07:28:18.6581422Z * [new branch] gh/bobrenjc93/551/orig -> origin/gh/bobrenjc93/551/orig 2025-09-07T07:28:18.6582855Z * [new branch] gh/bobrenjc93/552/base -> origin/gh/bobrenjc93/552/base 2025-09-07T07:28:18.6583615Z * [new branch] gh/bobrenjc93/552/head -> origin/gh/bobrenjc93/552/head 2025-09-07T07:28:18.6584435Z * [new branch] gh/bobrenjc93/552/orig -> origin/gh/bobrenjc93/552/orig 2025-09-07T07:28:18.6585577Z * [new branch] gh/bobrenjc93/553/base -> origin/gh/bobrenjc93/553/base 2025-09-07T07:28:18.6586445Z * [new branch] gh/bobrenjc93/553/head -> origin/gh/bobrenjc93/553/head 2025-09-07T07:28:18.6587283Z * [new branch] gh/bobrenjc93/553/orig -> origin/gh/bobrenjc93/553/orig 2025-09-07T07:28:18.6588443Z * [new branch] gh/bobrenjc93/554/base -> origin/gh/bobrenjc93/554/base 2025-09-07T07:28:18.6589426Z * [new branch] gh/bobrenjc93/554/head -> origin/gh/bobrenjc93/554/head 2025-09-07T07:28:18.6590226Z * [new branch] gh/bobrenjc93/554/orig -> origin/gh/bobrenjc93/554/orig 2025-09-07T07:28:18.6591479Z * [new branch] gh/bobrenjc93/555/base -> origin/gh/bobrenjc93/555/base 2025-09-07T07:28:18.6592245Z * [new branch] gh/bobrenjc93/555/head -> origin/gh/bobrenjc93/555/head 2025-09-07T07:28:18.6593118Z * [new branch] gh/bobrenjc93/555/orig -> origin/gh/bobrenjc93/555/orig 2025-09-07T07:28:18.6594289Z * [new branch] gh/bobrenjc93/556/base -> origin/gh/bobrenjc93/556/base 2025-09-07T07:28:18.6595207Z * [new branch] gh/bobrenjc93/556/head -> origin/gh/bobrenjc93/556/head 2025-09-07T07:28:18.6596010Z * [new branch] gh/bobrenjc93/556/orig -> origin/gh/bobrenjc93/556/orig 2025-09-07T07:28:18.6597468Z * [new branch] gh/briancoutinho/2/base -> origin/gh/briancoutinho/2/base 2025-09-07T07:28:18.6598511Z * [new branch] gh/briancoutinho/2/head -> origin/gh/briancoutinho/2/head 2025-09-07T07:28:18.6599997Z * [new branch] gh/c00w/23/base -> origin/gh/c00w/23/base 2025-09-07T07:28:18.6600953Z * [new branch] gh/c00w/23/head -> origin/gh/c00w/23/head 2025-09-07T07:28:18.6602168Z * [new branch] gh/c00w/48/base -> origin/gh/c00w/48/base 2025-09-07T07:28:18.6603086Z * [new branch] gh/c00w/48/head -> origin/gh/c00w/48/head 2025-09-07T07:28:18.6604042Z * [new branch] gh/c00w/48/orig -> origin/gh/c00w/48/orig 2025-09-07T07:28:18.6605281Z * [new branch] gh/c00w/53/base -> origin/gh/c00w/53/base 2025-09-07T07:28:18.6606105Z * [new branch] gh/c00w/53/head -> origin/gh/c00w/53/head 2025-09-07T07:28:18.6606992Z * [new branch] gh/c00w/53/orig -> origin/gh/c00w/53/orig 2025-09-07T07:28:18.6608281Z * [new branch] gh/c00w/54/base -> origin/gh/c00w/54/base 2025-09-07T07:28:18.6608874Z * [new branch] gh/c00w/54/head -> origin/gh/c00w/54/head 2025-09-07T07:28:18.6609787Z * [new branch] gh/c00w/54/orig -> origin/gh/c00w/54/orig 2025-09-07T07:28:18.6610990Z * [new branch] gh/c00w/55/base -> origin/gh/c00w/55/base 2025-09-07T07:28:18.6611986Z * [new branch] gh/c00w/55/head -> origin/gh/c00w/55/head 2025-09-07T07:28:18.6612925Z * [new branch] gh/c00w/55/orig -> origin/gh/c00w/55/orig 2025-09-07T07:28:18.6613994Z * [new branch] gh/c00w/56/base -> origin/gh/c00w/56/base 2025-09-07T07:28:18.6614968Z * [new branch] gh/c00w/56/head -> origin/gh/c00w/56/head 2025-09-07T07:28:18.6615888Z * [new branch] gh/c00w/56/orig -> origin/gh/c00w/56/orig 2025-09-07T07:28:18.6617333Z * [new branch] gh/clee2000/1/base -> origin/gh/clee2000/1/base 2025-09-07T07:28:18.6618415Z * [new branch] gh/clee2000/1/head -> origin/gh/clee2000/1/head 2025-09-07T07:28:18.6619167Z * [new branch] gh/clee2000/1/orig -> origin/gh/clee2000/1/orig 2025-09-07T07:28:18.6620675Z * [new branch] gh/coconutruben/1/base -> origin/gh/coconutruben/1/base 2025-09-07T07:28:18.6621706Z * [new branch] gh/coconutruben/1/head -> origin/gh/coconutruben/1/head 2025-09-07T07:28:18.6622998Z * [new branch] gh/coconutruben/11/base -> origin/gh/coconutruben/11/base 2025-09-07T07:28:18.6624011Z * [new branch] gh/coconutruben/11/head -> origin/gh/coconutruben/11/head 2025-09-07T07:28:18.6624942Z * [new branch] gh/coconutruben/11/orig -> origin/gh/coconutruben/11/orig 2025-09-07T07:28:18.6626605Z * [new branch] gh/coconutruben/12/base -> origin/gh/coconutruben/12/base 2025-09-07T07:28:18.6627723Z * [new branch] gh/coconutruben/12/head -> origin/gh/coconutruben/12/head 2025-09-07T07:28:18.6628834Z * [new branch] gh/coconutruben/12/orig -> origin/gh/coconutruben/12/orig 2025-09-07T07:28:18.6630141Z * [new branch] gh/coconutruben/13/base -> origin/gh/coconutruben/13/base 2025-09-07T07:28:18.6631096Z * [new branch] gh/coconutruben/13/head -> origin/gh/coconutruben/13/head 2025-09-07T07:28:18.6632000Z * [new branch] gh/coconutruben/13/orig -> origin/gh/coconutruben/13/orig 2025-09-07T07:28:18.6633319Z * [new branch] gh/coconutruben/14/base -> origin/gh/coconutruben/14/base 2025-09-07T07:28:18.6634266Z * [new branch] gh/coconutruben/14/head -> origin/gh/coconutruben/14/head 2025-09-07T07:28:18.6635187Z * [new branch] gh/coconutruben/14/orig -> origin/gh/coconutruben/14/orig 2025-09-07T07:28:18.6636764Z * [new branch] gh/coconutruben/15/base -> origin/gh/coconutruben/15/base 2025-09-07T07:28:18.6637814Z * [new branch] gh/coconutruben/15/head -> origin/gh/coconutruben/15/head 2025-09-07T07:28:18.6638795Z * [new branch] gh/coconutruben/15/orig -> origin/gh/coconutruben/15/orig 2025-09-07T07:28:18.6639958Z * [new branch] gh/coconutruben/16/base -> origin/gh/coconutruben/16/base 2025-09-07T07:28:18.6640803Z * [new branch] gh/coconutruben/16/head -> origin/gh/coconutruben/16/head 2025-09-07T07:28:18.6641648Z * [new branch] gh/coconutruben/16/orig -> origin/gh/coconutruben/16/orig 2025-09-07T07:28:18.6643137Z * [new branch] gh/coconutruben/17/base -> origin/gh/coconutruben/17/base 2025-09-07T07:28:18.6644088Z * [new branch] gh/coconutruben/17/head -> origin/gh/coconutruben/17/head 2025-09-07T07:28:18.6645036Z * [new branch] gh/coconutruben/17/orig -> origin/gh/coconutruben/17/orig 2025-09-07T07:28:18.6646313Z * [new branch] gh/coconutruben/18/base -> origin/gh/coconutruben/18/base 2025-09-07T07:28:18.6647294Z * [new branch] gh/coconutruben/18/head -> origin/gh/coconutruben/18/head 2025-09-07T07:28:18.6648148Z * [new branch] gh/coconutruben/18/orig -> origin/gh/coconutruben/18/orig 2025-09-07T07:28:18.6649372Z * [new branch] gh/coconutruben/19/base -> origin/gh/coconutruben/19/base 2025-09-07T07:28:18.6650433Z * [new branch] gh/coconutruben/19/head -> origin/gh/coconutruben/19/head 2025-09-07T07:28:18.6651364Z * [new branch] gh/coconutruben/19/orig -> origin/gh/coconutruben/19/orig 2025-09-07T07:28:18.6652753Z * [new branch] gh/coconutruben/20/base -> origin/gh/coconutruben/20/base 2025-09-07T07:28:18.6653645Z * [new branch] gh/coconutruben/20/head -> origin/gh/coconutruben/20/head 2025-09-07T07:28:18.6654614Z * [new branch] gh/coconutruben/20/orig -> origin/gh/coconutruben/20/orig 2025-09-07T07:28:18.6655903Z * [new branch] gh/coconutruben/21/base -> origin/gh/coconutruben/21/base 2025-09-07T07:28:18.6657030Z * [new branch] gh/coconutruben/21/head -> origin/gh/coconutruben/21/head 2025-09-07T07:28:18.6657665Z * [new branch] gh/coconutruben/21/orig -> origin/gh/coconutruben/21/orig 2025-09-07T07:28:18.6658891Z * [new branch] gh/coconutruben/22/base -> origin/gh/coconutruben/22/base 2025-09-07T07:28:18.6659793Z * [new branch] gh/coconutruben/22/head -> origin/gh/coconutruben/22/head 2025-09-07T07:28:18.6660782Z * [new branch] gh/coconutruben/22/orig -> origin/gh/coconutruben/22/orig 2025-09-07T07:28:18.6662076Z * [new branch] gh/coconutruben/24/base -> origin/gh/coconutruben/24/base 2025-09-07T07:28:18.6663164Z * [new branch] gh/coconutruben/24/head -> origin/gh/coconutruben/24/head 2025-09-07T07:28:18.6664070Z * [new branch] gh/coconutruben/24/orig -> origin/gh/coconutruben/24/orig 2025-09-07T07:28:18.6669904Z * [new branch] gh/coconutruben/25/base -> origin/gh/coconutruben/25/base 2025-09-07T07:28:18.6671496Z * [new branch] gh/coconutruben/25/head -> origin/gh/coconutruben/25/head 2025-09-07T07:28:18.6672639Z * [new branch] gh/coconutruben/25/orig -> origin/gh/coconutruben/25/orig 2025-09-07T07:28:18.6673965Z * [new branch] gh/coconutruben/28/base -> origin/gh/coconutruben/28/base 2025-09-07T07:28:18.6674915Z * [new branch] gh/coconutruben/28/head -> origin/gh/coconutruben/28/head 2025-09-07T07:28:18.6675806Z * [new branch] gh/coconutruben/28/orig -> origin/gh/coconutruben/28/orig 2025-09-07T07:28:18.6677152Z * [new branch] gh/coconutruben/29/base -> origin/gh/coconutruben/29/base 2025-09-07T07:28:18.6678076Z * [new branch] gh/coconutruben/29/head -> origin/gh/coconutruben/29/head 2025-09-07T07:28:18.6679072Z * [new branch] gh/coconutruben/29/orig -> origin/gh/coconutruben/29/orig 2025-09-07T07:28:18.6680339Z * [new branch] gh/coconutruben/30/base -> origin/gh/coconutruben/30/base 2025-09-07T07:28:18.6681338Z * [new branch] gh/coconutruben/30/head -> origin/gh/coconutruben/30/head 2025-09-07T07:28:18.6682296Z * [new branch] gh/coconutruben/30/orig -> origin/gh/coconutruben/30/orig 2025-09-07T07:28:18.6683986Z * [new branch] gh/coconutruben/31/base -> origin/gh/coconutruben/31/base 2025-09-07T07:28:18.6684957Z * [new branch] gh/coconutruben/31/head -> origin/gh/coconutruben/31/head 2025-09-07T07:28:18.6685873Z * [new branch] gh/coconutruben/31/orig -> origin/gh/coconutruben/31/orig 2025-09-07T07:28:18.6687368Z * [new branch] gh/coconutruben/32/base -> origin/gh/coconutruben/32/base 2025-09-07T07:28:18.6688357Z * [new branch] gh/coconutruben/32/head -> origin/gh/coconutruben/32/head 2025-09-07T07:28:18.6689295Z * [new branch] gh/coconutruben/32/orig -> origin/gh/coconutruben/32/orig 2025-09-07T07:28:18.6691035Z * [new branch] gh/coconutruben/33/base -> origin/gh/coconutruben/33/base 2025-09-07T07:28:18.6691994Z * [new branch] gh/coconutruben/33/head -> origin/gh/coconutruben/33/head 2025-09-07T07:28:18.6692953Z * [new branch] gh/coconutruben/33/orig -> origin/gh/coconutruben/33/orig 2025-09-07T07:28:18.6694115Z * [new branch] gh/coconutruben/34/base -> origin/gh/coconutruben/34/base 2025-09-07T07:28:18.6695040Z * [new branch] gh/coconutruben/34/head -> origin/gh/coconutruben/34/head 2025-09-07T07:28:18.6695907Z * [new branch] gh/coconutruben/34/orig -> origin/gh/coconutruben/34/orig 2025-09-07T07:28:18.6697150Z * [new branch] gh/coconutruben/35/base -> origin/gh/coconutruben/35/base 2025-09-07T07:28:18.6698016Z * [new branch] gh/coconutruben/35/head -> origin/gh/coconutruben/35/head 2025-09-07T07:28:18.6699109Z * [new branch] gh/coconutruben/35/orig -> origin/gh/coconutruben/35/orig 2025-09-07T07:28:18.6701579Z * [new branch] gh/coconutruben/36/base -> origin/gh/coconutruben/36/base 2025-09-07T07:28:18.6702852Z * [new branch] gh/coconutruben/36/head -> origin/gh/coconutruben/36/head 2025-09-07T07:28:18.6704464Z * [new branch] gh/coconutruben/36/orig -> origin/gh/coconutruben/36/orig 2025-09-07T07:28:18.6705986Z * [new branch] gh/coconutruben/37/base -> origin/gh/coconutruben/37/base 2025-09-07T07:28:18.6706828Z * [new branch] gh/coconutruben/37/head -> origin/gh/coconutruben/37/head 2025-09-07T07:28:18.6707676Z * [new branch] gh/coconutruben/37/orig -> origin/gh/coconutruben/37/orig 2025-09-07T07:28:18.6709027Z * [new branch] gh/coconutruben/38/base -> origin/gh/coconutruben/38/base 2025-09-07T07:28:18.6710113Z * [new branch] gh/coconutruben/38/head -> origin/gh/coconutruben/38/head 2025-09-07T07:28:18.6711108Z * [new branch] gh/coconutruben/38/orig -> origin/gh/coconutruben/38/orig 2025-09-07T07:28:18.6712447Z * [new branch] gh/coconutruben/39/base -> origin/gh/coconutruben/39/base 2025-09-07T07:28:18.6713311Z * [new branch] gh/coconutruben/39/head -> origin/gh/coconutruben/39/head 2025-09-07T07:28:18.6714476Z * [new branch] gh/coconutruben/39/orig -> origin/gh/coconutruben/39/orig 2025-09-07T07:28:18.6715959Z * [new branch] gh/coconutruben/40/base -> origin/gh/coconutruben/40/base 2025-09-07T07:28:18.6716795Z * [new branch] gh/coconutruben/40/head -> origin/gh/coconutruben/40/head 2025-09-07T07:28:18.6717710Z * [new branch] gh/coconutruben/40/orig -> origin/gh/coconutruben/40/orig 2025-09-07T07:28:18.6719254Z * [new branch] gh/coconutruben/41/base -> origin/gh/coconutruben/41/base 2025-09-07T07:28:18.6720215Z * [new branch] gh/coconutruben/41/head -> origin/gh/coconutruben/41/head 2025-09-07T07:28:18.6721076Z * [new branch] gh/coconutruben/41/orig -> origin/gh/coconutruben/41/orig 2025-09-07T07:28:18.6722495Z * [new branch] gh/coconutruben/42/base -> origin/gh/coconutruben/42/base 2025-09-07T07:28:18.6723516Z * [new branch] gh/coconutruben/42/head -> origin/gh/coconutruben/42/head 2025-09-07T07:28:18.6724479Z * [new branch] gh/coconutruben/42/orig -> origin/gh/coconutruben/42/orig 2025-09-07T07:28:18.6725900Z * [new branch] gh/coconutruben/43/base -> origin/gh/coconutruben/43/base 2025-09-07T07:28:18.6726876Z * [new branch] gh/coconutruben/43/head -> origin/gh/coconutruben/43/head 2025-09-07T07:28:18.6727797Z * [new branch] gh/coconutruben/43/orig -> origin/gh/coconutruben/43/orig 2025-09-07T07:28:18.6729253Z * [new branch] gh/coconutruben/44/base -> origin/gh/coconutruben/44/base 2025-09-07T07:28:18.6730231Z * [new branch] gh/coconutruben/44/head -> origin/gh/coconutruben/44/head 2025-09-07T07:28:18.6731192Z * [new branch] gh/coconutruben/44/orig -> origin/gh/coconutruben/44/orig 2025-09-07T07:28:18.6732571Z * [new branch] gh/coconutruben/45/base -> origin/gh/coconutruben/45/base 2025-09-07T07:28:18.6733496Z * [new branch] gh/coconutruben/45/head -> origin/gh/coconutruben/45/head 2025-09-07T07:28:18.6734427Z * [new branch] gh/coconutruben/45/orig -> origin/gh/coconutruben/45/orig 2025-09-07T07:28:18.6735696Z * [new branch] gh/coconutruben/46/base -> origin/gh/coconutruben/46/base 2025-09-07T07:28:18.6736655Z * [new branch] gh/coconutruben/46/head -> origin/gh/coconutruben/46/head 2025-09-07T07:28:18.6737599Z * [new branch] gh/coconutruben/46/orig -> origin/gh/coconutruben/46/orig 2025-09-07T07:28:18.6739081Z * [new branch] gh/coconutruben/47/base -> origin/gh/coconutruben/47/base 2025-09-07T07:28:18.6739974Z * [new branch] gh/coconutruben/47/head -> origin/gh/coconutruben/47/head 2025-09-07T07:28:18.6740894Z * [new branch] gh/coconutruben/47/orig -> origin/gh/coconutruben/47/orig 2025-09-07T07:28:18.6742256Z * [new branch] gh/coconutruben/48/base -> origin/gh/coconutruben/48/base 2025-09-07T07:28:18.6743213Z * [new branch] gh/coconutruben/48/head -> origin/gh/coconutruben/48/head 2025-09-07T07:28:18.6744075Z * [new branch] gh/coconutruben/48/orig -> origin/gh/coconutruben/48/orig 2025-09-07T07:28:18.6745714Z * [new branch] gh/coconutruben/49/base -> origin/gh/coconutruben/49/base 2025-09-07T07:28:18.6746582Z * [new branch] gh/coconutruben/49/head -> origin/gh/coconutruben/49/head 2025-09-07T07:28:18.6747546Z * [new branch] gh/coconutruben/49/orig -> origin/gh/coconutruben/49/orig 2025-09-07T07:28:18.6748874Z * [new branch] gh/coconutruben/50/base -> origin/gh/coconutruben/50/base 2025-09-07T07:28:18.6749872Z * [new branch] gh/coconutruben/50/head -> origin/gh/coconutruben/50/head 2025-09-07T07:28:18.6750815Z * [new branch] gh/coconutruben/50/orig -> origin/gh/coconutruben/50/orig 2025-09-07T07:28:18.6752094Z * [new branch] gh/coconutruben/51/base -> origin/gh/coconutruben/51/base 2025-09-07T07:28:18.6753015Z * [new branch] gh/coconutruben/51/head -> origin/gh/coconutruben/51/head 2025-09-07T07:28:18.6753890Z * [new branch] gh/coconutruben/51/orig -> origin/gh/coconutruben/51/orig 2025-09-07T07:28:18.6755333Z * [new branch] gh/coconutruben/52/base -> origin/gh/coconutruben/52/base 2025-09-07T07:28:18.6756321Z * [new branch] gh/coconutruben/52/head -> origin/gh/coconutruben/52/head 2025-09-07T07:28:18.6757286Z * [new branch] gh/coconutruben/52/orig -> origin/gh/coconutruben/52/orig 2025-09-07T07:28:18.6758697Z * [new branch] gh/coconutruben/53/base -> origin/gh/coconutruben/53/base 2025-09-07T07:28:18.6759549Z * [new branch] gh/coconutruben/53/head -> origin/gh/coconutruben/53/head 2025-09-07T07:28:18.6760467Z * [new branch] gh/coconutruben/53/orig -> origin/gh/coconutruben/53/orig 2025-09-07T07:28:18.6761672Z * [new branch] gh/coconutruben/54/base -> origin/gh/coconutruben/54/base 2025-09-07T07:28:18.6762736Z * [new branch] gh/coconutruben/54/head -> origin/gh/coconutruben/54/head 2025-09-07T07:28:18.6763629Z * [new branch] gh/coconutruben/54/orig -> origin/gh/coconutruben/54/orig 2025-09-07T07:28:18.6764929Z * [new branch] gh/coconutruben/55/base -> origin/gh/coconutruben/55/base 2025-09-07T07:28:18.6765882Z * [new branch] gh/coconutruben/55/head -> origin/gh/coconutruben/55/head 2025-09-07T07:28:18.6766876Z * [new branch] gh/coconutruben/55/orig -> origin/gh/coconutruben/55/orig 2025-09-07T07:28:18.6768225Z * [new branch] gh/coconutruben/56/base -> origin/gh/coconutruben/56/base 2025-09-07T07:28:18.6769161Z * [new branch] gh/coconutruben/56/head -> origin/gh/coconutruben/56/head 2025-09-07T07:28:18.6770137Z * [new branch] gh/coconutruben/56/orig -> origin/gh/coconutruben/56/orig 2025-09-07T07:28:18.6771450Z * [new branch] gh/coconutruben/57/base -> origin/gh/coconutruben/57/base 2025-09-07T07:28:18.6772493Z * [new branch] gh/coconutruben/57/head -> origin/gh/coconutruben/57/head 2025-09-07T07:28:18.6773426Z * [new branch] gh/coconutruben/57/orig -> origin/gh/coconutruben/57/orig 2025-09-07T07:28:18.6786889Z * [new branch] gh/coconutruben/58/base -> origin/gh/coconutruben/58/base 2025-09-07T07:28:18.6787478Z * [new branch] gh/coconutruben/58/head -> origin/gh/coconutruben/58/head 2025-09-07T07:28:18.6787719Z * [new branch] gh/coconutruben/58/orig -> origin/gh/coconutruben/58/orig 2025-09-07T07:28:18.6787929Z * [new branch] gh/coconutruben/59/base -> origin/gh/coconutruben/59/base 2025-09-07T07:28:18.6788131Z * [new branch] gh/coconutruben/59/head -> origin/gh/coconutruben/59/head 2025-09-07T07:28:18.6788331Z * [new branch] gh/coconutruben/59/orig -> origin/gh/coconutruben/59/orig 2025-09-07T07:28:18.6788534Z * [new branch] gh/coconutruben/60/base -> origin/gh/coconutruben/60/base 2025-09-07T07:28:18.6788734Z * [new branch] gh/coconutruben/60/head -> origin/gh/coconutruben/60/head 2025-09-07T07:28:18.6788944Z * [new branch] gh/coconutruben/60/orig -> origin/gh/coconutruben/60/orig 2025-09-07T07:28:18.6789143Z * [new branch] gh/coconutruben/61/base -> origin/gh/coconutruben/61/base 2025-09-07T07:28:18.6789347Z * [new branch] gh/coconutruben/61/head -> origin/gh/coconutruben/61/head 2025-09-07T07:28:18.6789552Z * [new branch] gh/coconutruben/61/orig -> origin/gh/coconutruben/61/orig 2025-09-07T07:28:18.6789754Z * [new branch] gh/coconutruben/62/base -> origin/gh/coconutruben/62/base 2025-09-07T07:28:18.6789955Z * [new branch] gh/coconutruben/62/head -> origin/gh/coconutruben/62/head 2025-09-07T07:28:18.6790154Z * [new branch] gh/coconutruben/62/orig -> origin/gh/coconutruben/62/orig 2025-09-07T07:28:18.6791241Z * [new branch] gh/coconutruben/63/base -> origin/gh/coconutruben/63/base 2025-09-07T07:28:18.6792160Z * [new branch] gh/coconutruben/63/head -> origin/gh/coconutruben/63/head 2025-09-07T07:28:18.6793051Z * [new branch] gh/coconutruben/63/orig -> origin/gh/coconutruben/63/orig 2025-09-07T07:28:18.6794294Z * [new branch] gh/coconutruben/64/base -> origin/gh/coconutruben/64/base 2025-09-07T07:28:18.6795272Z * [new branch] gh/coconutruben/64/head -> origin/gh/coconutruben/64/head 2025-09-07T07:28:18.6796191Z * [new branch] gh/coconutruben/64/orig -> origin/gh/coconutruben/64/orig 2025-09-07T07:28:18.6797476Z * [new branch] gh/coconutruben/65/base -> origin/gh/coconutruben/65/base 2025-09-07T07:28:18.6798403Z * [new branch] gh/coconutruben/65/head -> origin/gh/coconutruben/65/head 2025-09-07T07:28:18.6799297Z * [new branch] gh/coconutruben/65/orig -> origin/gh/coconutruben/65/orig 2025-09-07T07:28:18.6800636Z * [new branch] gh/coconutruben/66/base -> origin/gh/coconutruben/66/base 2025-09-07T07:28:18.6801540Z * [new branch] gh/coconutruben/66/head -> origin/gh/coconutruben/66/head 2025-09-07T07:28:18.6802406Z * [new branch] gh/coconutruben/66/orig -> origin/gh/coconutruben/66/orig 2025-09-07T07:28:18.6804575Z * [new branch] gh/codingwithsurya/12/base -> origin/gh/codingwithsurya/12/base 2025-09-07T07:28:18.6805494Z * [new branch] gh/codingwithsurya/12/head -> origin/gh/codingwithsurya/12/head 2025-09-07T07:28:18.6806487Z * [new branch] gh/codingwithsurya/12/orig -> origin/gh/codingwithsurya/12/orig 2025-09-07T07:28:18.6807638Z * [new branch] gh/codingwithsurya/14/base -> origin/gh/codingwithsurya/14/base 2025-09-07T07:28:18.6808693Z * [new branch] gh/codingwithsurya/14/head -> origin/gh/codingwithsurya/14/head 2025-09-07T07:28:18.6809384Z * [new branch] gh/codingwithsurya/14/orig -> origin/gh/codingwithsurya/14/orig 2025-09-07T07:28:18.6810757Z * [new branch] gh/codingwithsurya/15/base -> origin/gh/codingwithsurya/15/base 2025-09-07T07:28:18.6811715Z * [new branch] gh/codingwithsurya/15/head -> origin/gh/codingwithsurya/15/head 2025-09-07T07:28:18.6812708Z * [new branch] gh/codingwithsurya/15/orig -> origin/gh/codingwithsurya/15/orig 2025-09-07T07:28:18.6814083Z * [new branch] gh/codingwithsurya/16/base -> origin/gh/codingwithsurya/16/base 2025-09-07T07:28:18.6815030Z * [new branch] gh/codingwithsurya/16/head -> origin/gh/codingwithsurya/16/head 2025-09-07T07:28:18.6815879Z * [new branch] gh/codingwithsurya/16/orig -> origin/gh/codingwithsurya/16/orig 2025-09-07T07:28:18.6817259Z * [new branch] gh/codingwithsurya/17/base -> origin/gh/codingwithsurya/17/base 2025-09-07T07:28:18.6818261Z * [new branch] gh/codingwithsurya/17/head -> origin/gh/codingwithsurya/17/head 2025-09-07T07:28:18.6819076Z * [new branch] gh/codingwithsurya/17/orig -> origin/gh/codingwithsurya/17/orig 2025-09-07T07:28:18.6820385Z * [new branch] gh/codingwithsurya/18/base -> origin/gh/codingwithsurya/18/base 2025-09-07T07:28:18.6821307Z * [new branch] gh/codingwithsurya/18/head -> origin/gh/codingwithsurya/18/head 2025-09-07T07:28:18.6822172Z * [new branch] gh/codingwithsurya/18/orig -> origin/gh/codingwithsurya/18/orig 2025-09-07T07:28:18.6823527Z * [new branch] gh/codingwithsurya/19/base -> origin/gh/codingwithsurya/19/base 2025-09-07T07:28:18.6824447Z * [new branch] gh/codingwithsurya/19/head -> origin/gh/codingwithsurya/19/head 2025-09-07T07:28:18.6825324Z * [new branch] gh/codingwithsurya/19/orig -> origin/gh/codingwithsurya/19/orig 2025-09-07T07:28:18.6826606Z * [new branch] gh/codingwithsurya/20/base -> origin/gh/codingwithsurya/20/base 2025-09-07T07:28:18.6827497Z * [new branch] gh/codingwithsurya/20/head -> origin/gh/codingwithsurya/20/head 2025-09-07T07:28:18.6828364Z * [new branch] gh/codingwithsurya/20/orig -> origin/gh/codingwithsurya/20/orig 2025-09-07T07:28:18.6829717Z * [new branch] gh/codingwithsurya/21/base -> origin/gh/codingwithsurya/21/base 2025-09-07T07:28:18.6830647Z * [new branch] gh/codingwithsurya/21/head -> origin/gh/codingwithsurya/21/head 2025-09-07T07:28:18.6831538Z * [new branch] gh/codingwithsurya/21/orig -> origin/gh/codingwithsurya/21/orig 2025-09-07T07:28:18.6833032Z * [new branch] gh/colinchan15/1/base -> origin/gh/colinchan15/1/base 2025-09-07T07:28:18.6833920Z * [new branch] gh/colinchan15/1/head -> origin/gh/colinchan15/1/head 2025-09-07T07:28:18.6834936Z * [new branch] gh/colinchan15/2/base -> origin/gh/colinchan15/2/base 2025-09-07T07:28:18.6835754Z * [new branch] gh/colinchan15/2/head -> origin/gh/colinchan15/2/head 2025-09-07T07:28:18.6836836Z * [new branch] gh/colinchan15/3/base -> origin/gh/colinchan15/3/base 2025-09-07T07:28:18.6837654Z * [new branch] gh/colinchan15/3/head -> origin/gh/colinchan15/3/head 2025-09-07T07:28:18.6838756Z * [new branch] gh/colinchan15/6/base -> origin/gh/colinchan15/6/base 2025-09-07T07:28:18.6839580Z * [new branch] gh/colinchan15/6/head -> origin/gh/colinchan15/6/head 2025-09-07T07:28:18.6841148Z * [new branch] gh/davidberard98/382/base -> origin/gh/davidberard98/382/base 2025-09-07T07:28:18.6842144Z * [new branch] gh/davidberard98/382/head -> origin/gh/davidberard98/382/head 2025-09-07T07:28:18.6843073Z * [new branch] gh/davidberard98/382/orig -> origin/gh/davidberard98/382/orig 2025-09-07T07:28:18.6844252Z * [new branch] gh/davidberard98/386/base -> origin/gh/davidberard98/386/base 2025-09-07T07:28:18.6845279Z * [new branch] gh/davidberard98/386/head -> origin/gh/davidberard98/386/head 2025-09-07T07:28:18.6846131Z * [new branch] gh/davidberard98/386/orig -> origin/gh/davidberard98/386/orig 2025-09-07T07:28:18.6847408Z * [new branch] gh/davidberard98/391/base -> origin/gh/davidberard98/391/base 2025-09-07T07:28:18.6848322Z * [new branch] gh/davidberard98/391/head -> origin/gh/davidberard98/391/head 2025-09-07T07:28:18.6849156Z * [new branch] gh/davidberard98/391/orig -> origin/gh/davidberard98/391/orig 2025-09-07T07:28:18.6850350Z * [new branch] gh/davidberard98/392/base -> origin/gh/davidberard98/392/base 2025-09-07T07:28:18.6851216Z * [new branch] gh/davidberard98/392/head -> origin/gh/davidberard98/392/head 2025-09-07T07:28:18.6852103Z * [new branch] gh/davidberard98/392/orig -> origin/gh/davidberard98/392/orig 2025-09-07T07:28:18.6853408Z * [new branch] gh/davidberard98/394/base -> origin/gh/davidberard98/394/base 2025-09-07T07:28:18.6854335Z * [new branch] gh/davidberard98/394/head -> origin/gh/davidberard98/394/head 2025-09-07T07:28:18.6855213Z * [new branch] gh/davidberard98/394/orig -> origin/gh/davidberard98/394/orig 2025-09-07T07:28:18.6856449Z * [new branch] gh/davidberard98/396/base -> origin/gh/davidberard98/396/base 2025-09-07T07:28:18.6857297Z * [new branch] gh/davidberard98/396/head -> origin/gh/davidberard98/396/head 2025-09-07T07:28:18.6858175Z * [new branch] gh/davidberard98/396/orig -> origin/gh/davidberard98/396/orig 2025-09-07T07:28:18.6859577Z * [new branch] gh/davidberard98/397/base -> origin/gh/davidberard98/397/base 2025-09-07T07:28:18.6860417Z * [new branch] gh/davidberard98/397/head -> origin/gh/davidberard98/397/head 2025-09-07T07:28:18.6861424Z * [new branch] gh/davidberard98/397/orig -> origin/gh/davidberard98/397/orig 2025-09-07T07:28:18.6862630Z * [new branch] gh/davidberard98/398/base -> origin/gh/davidberard98/398/base 2025-09-07T07:28:18.6863433Z * [new branch] gh/davidberard98/398/head -> origin/gh/davidberard98/398/head 2025-09-07T07:28:18.6864301Z * [new branch] gh/davidberard98/398/orig -> origin/gh/davidberard98/398/orig 2025-09-07T07:28:18.6865551Z * [new branch] gh/davidberard98/399/base -> origin/gh/davidberard98/399/base 2025-09-07T07:28:18.6866554Z * [new branch] gh/davidberard98/399/head -> origin/gh/davidberard98/399/head 2025-09-07T07:28:18.6867371Z * [new branch] gh/davidberard98/399/orig -> origin/gh/davidberard98/399/orig 2025-09-07T07:28:18.6869068Z * [new branch] gh/davidberard98/400/base -> origin/gh/davidberard98/400/base 2025-09-07T07:28:18.6869594Z * [new branch] gh/davidberard98/400/head -> origin/gh/davidberard98/400/head 2025-09-07T07:28:18.6870446Z * [new branch] gh/davidberard98/400/orig -> origin/gh/davidberard98/400/orig 2025-09-07T07:28:18.6871581Z * [new branch] gh/davidberard98/401/base -> origin/gh/davidberard98/401/base 2025-09-07T07:28:18.6872458Z * [new branch] gh/davidberard98/401/head -> origin/gh/davidberard98/401/head 2025-09-07T07:28:18.6873383Z * [new branch] gh/davidberard98/401/orig -> origin/gh/davidberard98/401/orig 2025-09-07T07:28:18.6874559Z * [new branch] gh/davidberard98/402/base -> origin/gh/davidberard98/402/base 2025-09-07T07:28:18.6875420Z * [new branch] gh/davidberard98/402/head -> origin/gh/davidberard98/402/head 2025-09-07T07:28:18.6876353Z * [new branch] gh/davidberard98/402/orig -> origin/gh/davidberard98/402/orig 2025-09-07T07:28:18.6877600Z * [new branch] gh/davidberard98/403/base -> origin/gh/davidberard98/403/base 2025-09-07T07:28:18.6878428Z * [new branch] gh/davidberard98/403/head -> origin/gh/davidberard98/403/head 2025-09-07T07:28:18.6879277Z * [new branch] gh/davidberard98/403/orig -> origin/gh/davidberard98/403/orig 2025-09-07T07:28:18.6880604Z * [new branch] gh/davidberard98/404/base -> origin/gh/davidberard98/404/base 2025-09-07T07:28:18.6881483Z * [new branch] gh/davidberard98/404/head -> origin/gh/davidberard98/404/head 2025-09-07T07:28:18.6882213Z * [new branch] gh/davidberard98/404/orig -> origin/gh/davidberard98/404/orig 2025-09-07T07:28:18.6883442Z * [new branch] gh/davidberard98/405/base -> origin/gh/davidberard98/405/base 2025-09-07T07:28:18.6884341Z * [new branch] gh/davidberard98/405/head -> origin/gh/davidberard98/405/head 2025-09-07T07:28:18.6885306Z * [new branch] gh/davidberard98/405/orig -> origin/gh/davidberard98/405/orig 2025-09-07T07:28:18.6886569Z * [new branch] gh/davidberard98/406/base -> origin/gh/davidberard98/406/base 2025-09-07T07:28:18.6887585Z * [new branch] gh/davidberard98/406/head -> origin/gh/davidberard98/406/head 2025-09-07T07:28:18.6888515Z * [new branch] gh/davidberard98/406/orig -> origin/gh/davidberard98/406/orig 2025-09-07T07:28:18.6890320Z * [new branch] gh/davidberard98/407/base -> origin/gh/davidberard98/407/base 2025-09-07T07:28:18.6891187Z * [new branch] gh/davidberard98/407/head -> origin/gh/davidberard98/407/head 2025-09-07T07:28:18.6891996Z * [new branch] gh/davidberard98/407/orig -> origin/gh/davidberard98/407/orig 2025-09-07T07:28:18.6893210Z * [new branch] gh/davidberard98/408/base -> origin/gh/davidberard98/408/base 2025-09-07T07:28:18.6894082Z * [new branch] gh/davidberard98/408/head -> origin/gh/davidberard98/408/head 2025-09-07T07:28:18.6894924Z * [new branch] gh/davidberard98/408/orig -> origin/gh/davidberard98/408/orig 2025-09-07T07:28:18.6896078Z * [new branch] gh/davidberard98/409/base -> origin/gh/davidberard98/409/base 2025-09-07T07:28:18.6897067Z * [new branch] gh/davidberard98/409/head -> origin/gh/davidberard98/409/head 2025-09-07T07:28:18.6898083Z * [new branch] gh/davidberard98/409/orig -> origin/gh/davidberard98/409/orig 2025-09-07T07:28:18.6899583Z * [new branch] gh/desertfire/594/base -> origin/gh/desertfire/594/base 2025-09-07T07:28:18.6900374Z * [new branch] gh/desertfire/594/head -> origin/gh/desertfire/594/head 2025-09-07T07:28:18.6901293Z * [new branch] gh/desertfire/594/orig -> origin/gh/desertfire/594/orig 2025-09-07T07:28:18.6902474Z * [new branch] gh/desertfire/595/base -> origin/gh/desertfire/595/base 2025-09-07T07:28:18.6903325Z * [new branch] gh/desertfire/595/head -> origin/gh/desertfire/595/head 2025-09-07T07:28:18.6904157Z * [new branch] gh/desertfire/595/orig -> origin/gh/desertfire/595/orig 2025-09-07T07:28:18.6905381Z * [new branch] gh/desertfire/597/base -> origin/gh/desertfire/597/base 2025-09-07T07:28:18.6906255Z * [new branch] gh/desertfire/597/head -> origin/gh/desertfire/597/head 2025-09-07T07:28:18.6907130Z * [new branch] gh/desertfire/597/orig -> origin/gh/desertfire/597/orig 2025-09-07T07:28:18.6908693Z * [new branch] gh/dharakk/1/base -> origin/gh/dharakk/1/base 2025-09-07T07:28:18.6909616Z * [new branch] gh/dharakk/1/head -> origin/gh/dharakk/1/head 2025-09-07T07:28:18.6911044Z * [new branch] gh/drisspg/149/base -> origin/gh/drisspg/149/base 2025-09-07T07:28:18.6911938Z * [new branch] gh/drisspg/149/head -> origin/gh/drisspg/149/head 2025-09-07T07:28:18.6912805Z * [new branch] gh/drisspg/149/orig -> origin/gh/drisspg/149/orig 2025-09-07T07:28:18.6914044Z * [new branch] gh/drisspg/159/base -> origin/gh/drisspg/159/base 2025-09-07T07:28:18.6914911Z * [new branch] gh/drisspg/159/head -> origin/gh/drisspg/159/head 2025-09-07T07:28:18.6915746Z * [new branch] gh/drisspg/159/orig -> origin/gh/drisspg/159/orig 2025-09-07T07:28:18.6917080Z * [new branch] gh/drisspg/166/base -> origin/gh/drisspg/166/base 2025-09-07T07:28:18.6917834Z * [new branch] gh/drisspg/166/head -> origin/gh/drisspg/166/head 2025-09-07T07:28:18.6918720Z * [new branch] gh/drisspg/166/orig -> origin/gh/drisspg/166/orig 2025-09-07T07:28:18.6920175Z * [new branch] gh/drisspg/170/base -> origin/gh/drisspg/170/base 2025-09-07T07:28:18.6921055Z * [new branch] gh/drisspg/170/head -> origin/gh/drisspg/170/head 2025-09-07T07:28:18.6921923Z * [new branch] gh/drisspg/170/orig -> origin/gh/drisspg/170/orig 2025-09-07T07:28:18.6923116Z * [new branch] gh/drisspg/173/base -> origin/gh/drisspg/173/base 2025-09-07T07:28:18.6924030Z * [new branch] gh/drisspg/173/head -> origin/gh/drisspg/173/head 2025-09-07T07:28:18.6924842Z * [new branch] gh/drisspg/173/orig -> origin/gh/drisspg/173/orig 2025-09-07T07:28:18.6926014Z * [new branch] gh/drisspg/177/base -> origin/gh/drisspg/177/base 2025-09-07T07:28:18.6926889Z * [new branch] gh/drisspg/177/head -> origin/gh/drisspg/177/head 2025-09-07T07:28:18.6927709Z * [new branch] gh/drisspg/177/orig -> origin/gh/drisspg/177/orig 2025-09-07T07:28:18.6928883Z * [new branch] gh/drisspg/178/base -> origin/gh/drisspg/178/base 2025-09-07T07:28:18.6929739Z * [new branch] gh/drisspg/178/head -> origin/gh/drisspg/178/head 2025-09-07T07:28:18.6930493Z * [new branch] gh/drisspg/178/orig -> origin/gh/drisspg/178/orig 2025-09-07T07:28:18.6931720Z * [new branch] gh/drisspg/180/base -> origin/gh/drisspg/180/base 2025-09-07T07:28:18.6932594Z * [new branch] gh/drisspg/180/head -> origin/gh/drisspg/180/head 2025-09-07T07:28:18.6933399Z * [new branch] gh/drisspg/180/orig -> origin/gh/drisspg/180/orig 2025-09-07T07:28:18.6934607Z * [new branch] gh/drisspg/181/base -> origin/gh/drisspg/181/base 2025-09-07T07:28:18.6935466Z * [new branch] gh/drisspg/181/head -> origin/gh/drisspg/181/head 2025-09-07T07:28:18.6936343Z * [new branch] gh/drisspg/181/orig -> origin/gh/drisspg/181/orig 2025-09-07T07:28:18.6937532Z * [new branch] gh/drisspg/182/base -> origin/gh/drisspg/182/base 2025-09-07T07:28:18.6938387Z * [new branch] gh/drisspg/182/head -> origin/gh/drisspg/182/head 2025-09-07T07:28:18.6939503Z * [new branch] gh/drisspg/183/base -> origin/gh/drisspg/183/base 2025-09-07T07:28:18.6940398Z * [new branch] gh/drisspg/183/head -> origin/gh/drisspg/183/head 2025-09-07T07:28:18.6941378Z * [new branch] gh/drisspg/184/base -> origin/gh/drisspg/184/base 2025-09-07T07:28:18.6942111Z * [new branch] gh/drisspg/184/head -> origin/gh/drisspg/184/head 2025-09-07T07:28:18.6943433Z * [new branch] gh/drisspg/185/base -> origin/gh/drisspg/185/base 2025-09-07T07:28:18.6944289Z * [new branch] gh/drisspg/185/head -> origin/gh/drisspg/185/head 2025-09-07T07:28:18.6945485Z * [new branch] gh/drisspg/186/base -> origin/gh/drisspg/186/base 2025-09-07T07:28:18.6946331Z * [new branch] gh/drisspg/186/head -> origin/gh/drisspg/186/head 2025-09-07T07:28:18.6947175Z * [new branch] gh/drisspg/186/orig -> origin/gh/drisspg/186/orig 2025-09-07T07:28:18.6948321Z * [new branch] gh/drisspg/187/base -> origin/gh/drisspg/187/base 2025-09-07T07:28:18.6949147Z * [new branch] gh/drisspg/187/head -> origin/gh/drisspg/187/head 2025-09-07T07:28:18.6949980Z * [new branch] gh/drisspg/187/orig -> origin/gh/drisspg/187/orig 2025-09-07T07:28:18.6951209Z * [new branch] gh/drisspg/188/base -> origin/gh/drisspg/188/base 2025-09-07T07:28:18.6952100Z * [new branch] gh/drisspg/188/head -> origin/gh/drisspg/188/head 2025-09-07T07:28:18.6952862Z * [new branch] gh/drisspg/188/orig -> origin/gh/drisspg/188/orig 2025-09-07T07:28:18.6954465Z * [new branch] gh/drisspg/189/base -> origin/gh/drisspg/189/base 2025-09-07T07:28:18.6955346Z * [new branch] gh/drisspg/189/head -> origin/gh/drisspg/189/head 2025-09-07T07:28:18.6956263Z * [new branch] gh/drisspg/189/orig -> origin/gh/drisspg/189/orig 2025-09-07T07:28:18.6957527Z * [new branch] gh/drisspg/190/base -> origin/gh/drisspg/190/base 2025-09-07T07:28:18.6958368Z * [new branch] gh/drisspg/190/head -> origin/gh/drisspg/190/head 2025-09-07T07:28:18.6959193Z * [new branch] gh/drisspg/190/orig -> origin/gh/drisspg/190/orig 2025-09-07T07:28:18.6960401Z * [new branch] gh/drisspg/191/base -> origin/gh/drisspg/191/base 2025-09-07T07:28:18.6961275Z * [new branch] gh/drisspg/191/head -> origin/gh/drisspg/191/head 2025-09-07T07:28:18.6962126Z * [new branch] gh/drisspg/191/orig -> origin/gh/drisspg/191/orig 2025-09-07T07:28:18.6963365Z * [new branch] gh/drisspg/192/base -> origin/gh/drisspg/192/base 2025-09-07T07:28:18.6964205Z * [new branch] gh/drisspg/192/head -> origin/gh/drisspg/192/head 2025-09-07T07:28:18.6965049Z * [new branch] gh/drisspg/192/orig -> origin/gh/drisspg/192/orig 2025-09-07T07:28:18.6966394Z * [new branch] gh/drisspg/193/base -> origin/gh/drisspg/193/base 2025-09-07T07:28:18.6967253Z * [new branch] gh/drisspg/193/head -> origin/gh/drisspg/193/head 2025-09-07T07:28:18.6968147Z * [new branch] gh/drisspg/193/orig -> origin/gh/drisspg/193/orig 2025-09-07T07:28:18.6969262Z * [new branch] gh/drisspg/194/base -> origin/gh/drisspg/194/base 2025-09-07T07:28:18.6970163Z * [new branch] gh/drisspg/194/head -> origin/gh/drisspg/194/head 2025-09-07T07:28:18.6971092Z * [new branch] gh/drisspg/194/orig -> origin/gh/drisspg/194/orig 2025-09-07T07:28:18.6972339Z * [new branch] gh/drisspg/195/base -> origin/gh/drisspg/195/base 2025-09-07T07:28:18.6973173Z * [new branch] gh/drisspg/195/head -> origin/gh/drisspg/195/head 2025-09-07T07:28:18.6974021Z * [new branch] gh/drisspg/195/orig -> origin/gh/drisspg/195/orig 2025-09-07T07:28:18.6975227Z * [new branch] gh/drisspg/196/base -> origin/gh/drisspg/196/base 2025-09-07T07:28:18.6976042Z * [new branch] gh/drisspg/196/head -> origin/gh/drisspg/196/head 2025-09-07T07:28:18.6976913Z * [new branch] gh/drisspg/196/orig -> origin/gh/drisspg/196/orig 2025-09-07T07:28:18.6978200Z * [new branch] gh/drisspg/197/base -> origin/gh/drisspg/197/base 2025-09-07T07:28:18.6979014Z * [new branch] gh/drisspg/197/head -> origin/gh/drisspg/197/head 2025-09-07T07:28:18.6979863Z * [new branch] gh/drisspg/197/orig -> origin/gh/drisspg/197/orig 2025-09-07T07:28:18.6981081Z * [new branch] gh/drisspg/198/base -> origin/gh/drisspg/198/base 2025-09-07T07:28:18.6981899Z * [new branch] gh/drisspg/198/head -> origin/gh/drisspg/198/head 2025-09-07T07:28:18.6982743Z * [new branch] gh/drisspg/198/orig -> origin/gh/drisspg/198/orig 2025-09-07T07:28:18.6983981Z * [new branch] gh/drisspg/199/base -> origin/gh/drisspg/199/base 2025-09-07T07:28:18.6984828Z * [new branch] gh/drisspg/199/head -> origin/gh/drisspg/199/head 2025-09-07T07:28:18.6985699Z * [new branch] gh/drisspg/199/orig -> origin/gh/drisspg/199/orig 2025-09-07T07:28:18.6987356Z * [new branch] gh/dsjohns2/1/base -> origin/gh/dsjohns2/1/base 2025-09-07T07:28:18.6988119Z * [new branch] gh/dsjohns2/1/head -> origin/gh/dsjohns2/1/head 2025-09-07T07:28:18.6989687Z * [new branch] gh/eellison/784/base -> origin/gh/eellison/784/base 2025-09-07T07:28:18.6990543Z * [new branch] gh/eellison/784/head -> origin/gh/eellison/784/head 2025-09-07T07:28:18.6991416Z * [new branch] gh/eellison/784/orig -> origin/gh/eellison/784/orig 2025-09-07T07:28:18.6992826Z * [new branch] gh/eellison/785/base -> origin/gh/eellison/785/base 2025-09-07T07:28:18.6993695Z * [new branch] gh/eellison/785/head -> origin/gh/eellison/785/head 2025-09-07T07:28:18.6994521Z * [new branch] gh/eellison/785/orig -> origin/gh/eellison/785/orig 2025-09-07T07:28:18.6995747Z * [new branch] gh/eellison/789/base -> origin/gh/eellison/789/base 2025-09-07T07:28:18.6996613Z * [new branch] gh/eellison/789/head -> origin/gh/eellison/789/head 2025-09-07T07:28:18.6997477Z * [new branch] gh/eellison/789/orig -> origin/gh/eellison/789/orig 2025-09-07T07:28:18.6998624Z * [new branch] gh/eellison/800/base -> origin/gh/eellison/800/base 2025-09-07T07:28:18.6999504Z * [new branch] gh/eellison/800/head -> origin/gh/eellison/800/head 2025-09-07T07:28:18.7000347Z * [new branch] gh/eellison/800/orig -> origin/gh/eellison/800/orig 2025-09-07T07:28:18.7001546Z * [new branch] gh/eellison/801/base -> origin/gh/eellison/801/base 2025-09-07T07:28:18.7002472Z * [new branch] gh/eellison/801/head -> origin/gh/eellison/801/head 2025-09-07T07:28:18.7003386Z * [new branch] gh/eellison/801/orig -> origin/gh/eellison/801/orig 2025-09-07T07:28:18.7004601Z * [new branch] gh/eellison/802/base -> origin/gh/eellison/802/base 2025-09-07T07:28:18.7005463Z * [new branch] gh/eellison/802/head -> origin/gh/eellison/802/head 2025-09-07T07:28:18.7006301Z * [new branch] gh/eellison/802/orig -> origin/gh/eellison/802/orig 2025-09-07T07:28:18.7007410Z * [new branch] gh/eellison/805/base -> origin/gh/eellison/805/base 2025-09-07T07:28:18.7008257Z * [new branch] gh/eellison/805/head -> origin/gh/eellison/805/head 2025-09-07T07:28:18.7009132Z * [new branch] gh/eellison/805/orig -> origin/gh/eellison/805/orig 2025-09-07T07:28:18.7010491Z * [new branch] gh/eellison/808/base -> origin/gh/eellison/808/base 2025-09-07T07:28:18.7011407Z * [new branch] gh/eellison/808/head -> origin/gh/eellison/808/head 2025-09-07T07:28:18.7012241Z * [new branch] gh/eellison/808/orig -> origin/gh/eellison/808/orig 2025-09-07T07:28:18.7013702Z * [new branch] gh/eellison/809/base -> origin/gh/eellison/809/base 2025-09-07T07:28:18.7014616Z * [new branch] gh/eellison/809/head -> origin/gh/eellison/809/head 2025-09-07T07:28:18.7015491Z * [new branch] gh/eellison/809/orig -> origin/gh/eellison/809/orig 2025-09-07T07:28:18.7016677Z * [new branch] gh/eellison/813/base -> origin/gh/eellison/813/base 2025-09-07T07:28:18.7017597Z * [new branch] gh/eellison/813/head -> origin/gh/eellison/813/head 2025-09-07T07:28:18.7018489Z * [new branch] gh/eellison/813/orig -> origin/gh/eellison/813/orig 2025-09-07T07:28:18.7019673Z * [new branch] gh/eellison/814/base -> origin/gh/eellison/814/base 2025-09-07T07:28:18.7020554Z * [new branch] gh/eellison/814/head -> origin/gh/eellison/814/head 2025-09-07T07:28:18.7021427Z * [new branch] gh/eellison/814/orig -> origin/gh/eellison/814/orig 2025-09-07T07:28:18.7023025Z * [new branch] gh/eellison/815/base -> origin/gh/eellison/815/base 2025-09-07T07:28:18.7023780Z * [new branch] gh/eellison/815/head -> origin/gh/eellison/815/head 2025-09-07T07:28:18.7024662Z * [new branch] gh/eellison/815/orig -> origin/gh/eellison/815/orig 2025-09-07T07:28:18.7026027Z * [new branch] gh/eellison/816/base -> origin/gh/eellison/816/base 2025-09-07T07:28:18.7026718Z * [new branch] gh/eellison/816/head -> origin/gh/eellison/816/head 2025-09-07T07:28:18.7027545Z * [new branch] gh/eellison/816/orig -> origin/gh/eellison/816/orig 2025-09-07T07:28:18.7028670Z * [new branch] gh/eellison/817/base -> origin/gh/eellison/817/base 2025-09-07T07:28:18.7029552Z * [new branch] gh/eellison/817/head -> origin/gh/eellison/817/head 2025-09-07T07:28:18.7030361Z * [new branch] gh/eellison/817/orig -> origin/gh/eellison/817/orig 2025-09-07T07:28:18.7031623Z * [new branch] gh/eellison/818/base -> origin/gh/eellison/818/base 2025-09-07T07:28:18.7032531Z * [new branch] gh/eellison/818/head -> origin/gh/eellison/818/head 2025-09-07T07:28:18.7033456Z * [new branch] gh/eellison/818/orig -> origin/gh/eellison/818/orig 2025-09-07T07:28:18.7034821Z * [new branch] gh/eellison/819/base -> origin/gh/eellison/819/base 2025-09-07T07:28:18.7035654Z * [new branch] gh/eellison/819/head -> origin/gh/eellison/819/head 2025-09-07T07:28:18.7036501Z * [new branch] gh/eellison/819/orig -> origin/gh/eellison/819/orig 2025-09-07T07:28:18.7038355Z * [new branch] gh/eellison/820/base -> origin/gh/eellison/820/base 2025-09-07T07:28:18.7039370Z * [new branch] gh/eellison/820/head -> origin/gh/eellison/820/head 2025-09-07T07:28:18.7040235Z * [new branch] gh/eellison/820/orig -> origin/gh/eellison/820/orig 2025-09-07T07:28:18.7041431Z * [new branch] gh/eellison/821/base -> origin/gh/eellison/821/base 2025-09-07T07:28:18.7042296Z * [new branch] gh/eellison/821/head -> origin/gh/eellison/821/head 2025-09-07T07:28:18.7043218Z * [new branch] gh/eellison/821/orig -> origin/gh/eellison/821/orig 2025-09-07T07:28:18.7044469Z * [new branch] gh/eellison/822/base -> origin/gh/eellison/822/base 2025-09-07T07:28:18.7045369Z * [new branch] gh/eellison/822/head -> origin/gh/eellison/822/head 2025-09-07T07:28:18.7046193Z * [new branch] gh/eellison/822/orig -> origin/gh/eellison/822/orig 2025-09-07T07:28:18.7047432Z * [new branch] gh/eellison/823/base -> origin/gh/eellison/823/base 2025-09-07T07:28:18.7048333Z * [new branch] gh/eellison/823/head -> origin/gh/eellison/823/head 2025-09-07T07:28:18.7049150Z * [new branch] gh/eellison/823/orig -> origin/gh/eellison/823/orig 2025-09-07T07:28:18.7050676Z * [new branch] gh/etaf/132/base -> origin/gh/etaf/132/base 2025-09-07T07:28:18.7051572Z * [new branch] gh/etaf/132/head -> origin/gh/etaf/132/head 2025-09-07T07:28:18.7052376Z * [new branch] gh/etaf/132/orig -> origin/gh/etaf/132/orig 2025-09-07T07:28:18.7053574Z * [new branch] gh/etaf/138/base -> origin/gh/etaf/138/base 2025-09-07T07:28:18.7054384Z * [new branch] gh/etaf/138/head -> origin/gh/etaf/138/head 2025-09-07T07:28:18.7055241Z * [new branch] gh/etaf/138/orig -> origin/gh/etaf/138/orig 2025-09-07T07:28:18.7056496Z * [new branch] gh/etaf/140/base -> origin/gh/etaf/140/base 2025-09-07T07:28:18.7057375Z * [new branch] gh/etaf/140/head -> origin/gh/etaf/140/head 2025-09-07T07:28:18.7058352Z * [new branch] gh/etaf/140/orig -> origin/gh/etaf/140/orig 2025-09-07T07:28:18.7059503Z * [new branch] gh/etaf/143/base -> origin/gh/etaf/143/base 2025-09-07T07:28:18.7060303Z * [new branch] gh/etaf/143/head -> origin/gh/etaf/143/head 2025-09-07T07:28:18.7061137Z * [new branch] gh/etaf/143/orig -> origin/gh/etaf/143/orig 2025-09-07T07:28:18.7062412Z * [new branch] gh/etaf/147/base -> origin/gh/etaf/147/base 2025-09-07T07:28:18.7063302Z * [new branch] gh/etaf/147/head -> origin/gh/etaf/147/head 2025-09-07T07:28:18.7064532Z * [new branch] gh/etaf/151/base -> origin/gh/etaf/151/base 2025-09-07T07:28:18.7065628Z * [new branch] gh/etaf/151/head -> origin/gh/etaf/151/head 2025-09-07T07:28:18.7066560Z * [new branch] gh/etaf/151/orig -> origin/gh/etaf/151/orig 2025-09-07T07:28:18.7067872Z * [new branch] gh/etaf/152/base -> origin/gh/etaf/152/base 2025-09-07T07:28:18.7068791Z * [new branch] gh/etaf/152/head -> origin/gh/etaf/152/head 2025-09-07T07:28:18.7069647Z * [new branch] gh/etaf/152/orig -> origin/gh/etaf/152/orig 2025-09-07T07:28:18.7070955Z * [new branch] gh/etaf/153/base -> origin/gh/etaf/153/base 2025-09-07T07:28:18.7071841Z * [new branch] gh/etaf/153/head -> origin/gh/etaf/153/head 2025-09-07T07:28:18.7072665Z * [new branch] gh/etaf/153/orig -> origin/gh/etaf/153/orig 2025-09-07T07:28:18.7074063Z * [new branch] gh/etaf/154/base -> origin/gh/etaf/154/base 2025-09-07T07:28:18.7075017Z * [new branch] gh/etaf/154/head -> origin/gh/etaf/154/head 2025-09-07T07:28:18.7075820Z * [new branch] gh/etaf/154/orig -> origin/gh/etaf/154/orig 2025-09-07T07:28:18.7077028Z * [new branch] gh/etaf/155/base -> origin/gh/etaf/155/base 2025-09-07T07:28:18.7077943Z * [new branch] gh/etaf/155/head -> origin/gh/etaf/155/head 2025-09-07T07:28:18.7078846Z * [new branch] gh/etaf/155/orig -> origin/gh/etaf/155/orig 2025-09-07T07:28:18.7079929Z * [new branch] gh/etaf/156/base -> origin/gh/etaf/156/base 2025-09-07T07:28:18.7080820Z * [new branch] gh/etaf/156/head -> origin/gh/etaf/156/head 2025-09-07T07:28:18.7081701Z * [new branch] gh/etaf/156/orig -> origin/gh/etaf/156/orig 2025-09-07T07:28:18.7083064Z * [new branch] gh/etaf/157/base -> origin/gh/etaf/157/base 2025-09-07T07:28:18.7083962Z * [new branch] gh/etaf/157/head -> origin/gh/etaf/157/head 2025-09-07T07:28:18.7084862Z * [new branch] gh/etaf/157/orig -> origin/gh/etaf/157/orig 2025-09-07T07:28:18.7086012Z * [new branch] gh/etaf/158/base -> origin/gh/etaf/158/base 2025-09-07T07:28:18.7086948Z * [new branch] gh/etaf/158/head -> origin/gh/etaf/158/head 2025-09-07T07:28:18.7087809Z * [new branch] gh/etaf/158/orig -> origin/gh/etaf/158/orig 2025-09-07T07:28:18.7089013Z * [new branch] gh/etaf/159/base -> origin/gh/etaf/159/base 2025-09-07T07:28:18.7089968Z * [new branch] gh/etaf/159/head -> origin/gh/etaf/159/head 2025-09-07T07:28:18.7090809Z * [new branch] gh/etaf/159/orig -> origin/gh/etaf/159/orig 2025-09-07T07:28:18.7092043Z * [new branch] gh/etaf/160/base -> origin/gh/etaf/160/base 2025-09-07T07:28:18.7092986Z * [new branch] gh/etaf/160/head -> origin/gh/etaf/160/head 2025-09-07T07:28:18.7093870Z * [new branch] gh/etaf/160/orig -> origin/gh/etaf/160/orig 2025-09-07T07:28:18.7095107Z * [new branch] gh/etaf/161/base -> origin/gh/etaf/161/base 2025-09-07T07:28:18.7096116Z * [new branch] gh/etaf/161/head -> origin/gh/etaf/161/head 2025-09-07T07:28:18.7096938Z * [new branch] gh/etaf/161/orig -> origin/gh/etaf/161/orig 2025-09-07T07:28:18.7098474Z * [new branch] gh/etaf/162/base -> origin/gh/etaf/162/base 2025-09-07T07:28:18.7099348Z * [new branch] gh/etaf/162/head -> origin/gh/etaf/162/head 2025-09-07T07:28:18.7100272Z * [new branch] gh/etaf/162/orig -> origin/gh/etaf/162/orig 2025-09-07T07:28:18.7101565Z * [new branch] gh/etaf/163/base -> origin/gh/etaf/163/base 2025-09-07T07:28:18.7102386Z * [new branch] gh/etaf/163/head -> origin/gh/etaf/163/head 2025-09-07T07:28:18.7103181Z * [new branch] gh/etaf/163/orig -> origin/gh/etaf/163/orig 2025-09-07T07:28:18.7104498Z * [new branch] gh/etaf/164/base -> origin/gh/etaf/164/base 2025-09-07T07:28:18.7105452Z * [new branch] gh/etaf/164/head -> origin/gh/etaf/164/head 2025-09-07T07:28:18.7106305Z * [new branch] gh/etaf/164/orig -> origin/gh/etaf/164/orig 2025-09-07T07:28:18.7107562Z * [new branch] gh/etaf/165/base -> origin/gh/etaf/165/base 2025-09-07T07:28:18.7108415Z * [new branch] gh/etaf/165/orig -> origin/gh/etaf/165/orig 2025-09-07T07:28:18.7109668Z * [new branch] gh/etaf/166/base -> origin/gh/etaf/166/base 2025-09-07T07:28:18.7110563Z * [new branch] gh/etaf/166/head -> origin/gh/etaf/166/head 2025-09-07T07:28:18.7111455Z * [new branch] gh/etaf/166/orig -> origin/gh/etaf/166/orig 2025-09-07T07:28:18.7112746Z * [new branch] gh/etaf/167/base -> origin/gh/etaf/167/base 2025-09-07T07:28:18.7113643Z * [new branch] gh/etaf/167/head -> origin/gh/etaf/167/head 2025-09-07T07:28:18.7114487Z * [new branch] gh/etaf/167/orig -> origin/gh/etaf/167/orig 2025-09-07T07:28:18.7115753Z * [new branch] gh/etaf/168/base -> origin/gh/etaf/168/base 2025-09-07T07:28:18.7116706Z * [new branch] gh/etaf/168/head -> origin/gh/etaf/168/head 2025-09-07T07:28:18.7117602Z * [new branch] gh/etaf/168/orig -> origin/gh/etaf/168/orig 2025-09-07T07:28:18.7118910Z * [new branch] gh/etaf/169/base -> origin/gh/etaf/169/base 2025-09-07T07:28:18.7119764Z * [new branch] gh/etaf/169/head -> origin/gh/etaf/169/head 2025-09-07T07:28:18.7120591Z * [new branch] gh/etaf/169/orig -> origin/gh/etaf/169/orig 2025-09-07T07:28:18.7122113Z * [new branch] gh/exclamaforte/1/base -> origin/gh/exclamaforte/1/base 2025-09-07T07:28:18.7122956Z * [new branch] gh/exclamaforte/1/head -> origin/gh/exclamaforte/1/head 2025-09-07T07:28:18.7124188Z * [new branch] gh/exclamaforte/2/base -> origin/gh/exclamaforte/2/base 2025-09-07T07:28:18.7124977Z * [new branch] gh/exclamaforte/2/head -> origin/gh/exclamaforte/2/head 2025-09-07T07:28:18.7126282Z * [new branch] gh/exclamaforte/3/base -> origin/gh/exclamaforte/3/base 2025-09-07T07:28:18.7127069Z * [new branch] gh/exclamaforte/3/head -> origin/gh/exclamaforte/3/head 2025-09-07T07:28:18.7128290Z * [new branch] gh/exclamaforte/4/base -> origin/gh/exclamaforte/4/base 2025-09-07T07:28:18.7129159Z * [new branch] gh/exclamaforte/4/head -> origin/gh/exclamaforte/4/head 2025-09-07T07:28:18.7130722Z * [new branch] gh/ezyang/2374/base -> origin/gh/ezyang/2374/base 2025-09-07T07:28:18.7131580Z * [new branch] gh/ezyang/2374/head -> origin/gh/ezyang/2374/head 2025-09-07T07:28:18.7132781Z * [new branch] gh/ezyang/2374/orig -> origin/gh/ezyang/2374/orig 2025-09-07T07:28:18.7133618Z * [new branch] gh/ezyang/2973/base -> origin/gh/ezyang/2973/base 2025-09-07T07:28:18.7134557Z * [new branch] gh/ezyang/2973/head -> origin/gh/ezyang/2973/head 2025-09-07T07:28:18.7135500Z * [new branch] gh/ezyang/2973/orig -> origin/gh/ezyang/2973/orig 2025-09-07T07:28:18.7136588Z * [new branch] gh/ezyang/2974/base -> origin/gh/ezyang/2974/base 2025-09-07T07:28:18.7137448Z * [new branch] gh/ezyang/2974/head -> origin/gh/ezyang/2974/head 2025-09-07T07:28:18.7138342Z * [new branch] gh/ezyang/2974/orig -> origin/gh/ezyang/2974/orig 2025-09-07T07:28:18.7139516Z * [new branch] gh/ezyang/3074/base -> origin/gh/ezyang/3074/base 2025-09-07T07:28:18.7140327Z * [new branch] gh/ezyang/3074/head -> origin/gh/ezyang/3074/head 2025-09-07T07:28:18.7141217Z * [new branch] gh/ezyang/3074/orig -> origin/gh/ezyang/3074/orig 2025-09-07T07:28:18.7142376Z * [new branch] gh/ezyang/3088/base -> origin/gh/ezyang/3088/base 2025-09-07T07:28:18.7143238Z * [new branch] gh/ezyang/3088/head -> origin/gh/ezyang/3088/head 2025-09-07T07:28:18.7144147Z * [new branch] gh/ezyang/3088/orig -> origin/gh/ezyang/3088/orig 2025-09-07T07:28:18.7145295Z * [new branch] gh/ezyang/3092/base -> origin/gh/ezyang/3092/base 2025-09-07T07:28:18.7146277Z * [new branch] gh/ezyang/3092/head -> origin/gh/ezyang/3092/head 2025-09-07T07:28:18.7147164Z * [new branch] gh/ezyang/3092/orig -> origin/gh/ezyang/3092/orig 2025-09-07T07:28:18.7148316Z * [new branch] gh/ezyang/3103/base -> origin/gh/ezyang/3103/base 2025-09-07T07:28:18.7149168Z * [new branch] gh/ezyang/3103/head -> origin/gh/ezyang/3103/head 2025-09-07T07:28:18.7150055Z * [new branch] gh/ezyang/3103/orig -> origin/gh/ezyang/3103/orig 2025-09-07T07:28:18.7151192Z * [new branch] gh/ezyang/3105/base -> origin/gh/ezyang/3105/base 2025-09-07T07:28:18.7152016Z * [new branch] gh/ezyang/3105/head -> origin/gh/ezyang/3105/head 2025-09-07T07:28:18.7152889Z * [new branch] gh/ezyang/3105/orig -> origin/gh/ezyang/3105/orig 2025-09-07T07:28:18.7154057Z * [new branch] gh/ezyang/3114/base -> origin/gh/ezyang/3114/base 2025-09-07T07:28:18.7154996Z * [new branch] gh/ezyang/3114/head -> origin/gh/ezyang/3114/head 2025-09-07T07:28:18.7155847Z * [new branch] gh/ezyang/3114/orig -> origin/gh/ezyang/3114/orig 2025-09-07T07:28:18.7157002Z * [new branch] gh/ezyang/3116/base -> origin/gh/ezyang/3116/base 2025-09-07T07:28:18.7157923Z * [new branch] gh/ezyang/3116/head -> origin/gh/ezyang/3116/head 2025-09-07T07:28:18.7158746Z * [new branch] gh/ezyang/3116/orig -> origin/gh/ezyang/3116/orig 2025-09-07T07:28:18.7159961Z * [new branch] gh/ezyang/3120/base -> origin/gh/ezyang/3120/base 2025-09-07T07:28:18.7160835Z * [new branch] gh/ezyang/3120/head -> origin/gh/ezyang/3120/head 2025-09-07T07:28:18.7161712Z * [new branch] gh/ezyang/3120/orig -> origin/gh/ezyang/3120/orig 2025-09-07T07:28:18.7162851Z * [new branch] gh/ezyang/3122/base -> origin/gh/ezyang/3122/base 2025-09-07T07:28:18.7163704Z * [new branch] gh/ezyang/3122/head -> origin/gh/ezyang/3122/head 2025-09-07T07:28:18.7164530Z * [new branch] gh/ezyang/3122/orig -> origin/gh/ezyang/3122/orig 2025-09-07T07:28:18.7166501Z * [new branch] gh/ezyang/3123/base -> origin/gh/ezyang/3123/base 2025-09-07T07:28:18.7167383Z * [new branch] gh/ezyang/3123/head -> origin/gh/ezyang/3123/head 2025-09-07T07:28:18.7168903Z * [new branch] gh/ezyang/3123/orig -> origin/gh/ezyang/3123/orig 2025-09-07T07:28:18.7169538Z * [new branch] gh/ezyang/3125/base -> origin/gh/ezyang/3125/base 2025-09-07T07:28:18.7170427Z * [new branch] gh/ezyang/3125/head -> origin/gh/ezyang/3125/head 2025-09-07T07:28:18.7171237Z * [new branch] gh/ezyang/3125/orig -> origin/gh/ezyang/3125/orig 2025-09-07T07:28:18.7172443Z * [new branch] gh/ezyang/3126/base -> origin/gh/ezyang/3126/base 2025-09-07T07:28:18.7173261Z * [new branch] gh/ezyang/3126/head -> origin/gh/ezyang/3126/head 2025-09-07T07:28:18.7174112Z * [new branch] gh/ezyang/3126/orig -> origin/gh/ezyang/3126/orig 2025-09-07T07:28:18.7175591Z * [new branch] gh/ezyang/3127/base -> origin/gh/ezyang/3127/base 2025-09-07T07:28:18.7176470Z * [new branch] gh/ezyang/3127/head -> origin/gh/ezyang/3127/head 2025-09-07T07:28:18.7177361Z * [new branch] gh/ezyang/3127/orig -> origin/gh/ezyang/3127/orig 2025-09-07T07:28:18.7178630Z * [new branch] gh/ezyang/3128/base -> origin/gh/ezyang/3128/base 2025-09-07T07:28:18.7179476Z * [new branch] gh/ezyang/3128/head -> origin/gh/ezyang/3128/head 2025-09-07T07:28:18.7180359Z * [new branch] gh/ezyang/3128/orig -> origin/gh/ezyang/3128/orig 2025-09-07T07:28:18.7181556Z * [new branch] gh/ezyang/3129/base -> origin/gh/ezyang/3129/base 2025-09-07T07:28:18.7182479Z * [new branch] gh/ezyang/3129/head -> origin/gh/ezyang/3129/head 2025-09-07T07:28:18.7183340Z * [new branch] gh/ezyang/3129/orig -> origin/gh/ezyang/3129/orig 2025-09-07T07:28:18.7184510Z * [new branch] gh/ezyang/3130/base -> origin/gh/ezyang/3130/base 2025-09-07T07:28:18.7185379Z * [new branch] gh/ezyang/3130/head -> origin/gh/ezyang/3130/head 2025-09-07T07:28:18.7186209Z * [new branch] gh/ezyang/3130/orig -> origin/gh/ezyang/3130/orig 2025-09-07T07:28:18.7187410Z * [new branch] gh/ezyang/3131/base -> origin/gh/ezyang/3131/base 2025-09-07T07:28:18.7188247Z * [new branch] gh/ezyang/3131/head -> origin/gh/ezyang/3131/head 2025-09-07T07:28:18.7189068Z * [new branch] gh/ezyang/3131/orig -> origin/gh/ezyang/3131/orig 2025-09-07T07:28:18.7190281Z * [new branch] gh/ezyang/3132/base -> origin/gh/ezyang/3132/base 2025-09-07T07:28:18.7191163Z * [new branch] gh/ezyang/3132/head -> origin/gh/ezyang/3132/head 2025-09-07T07:28:18.7192044Z * [new branch] gh/ezyang/3132/orig -> origin/gh/ezyang/3132/orig 2025-09-07T07:28:18.7193257Z * [new branch] gh/ezyang/3133/base -> origin/gh/ezyang/3133/base 2025-09-07T07:28:18.7194133Z * [new branch] gh/ezyang/3133/head -> origin/gh/ezyang/3133/head 2025-09-07T07:28:18.7194995Z * [new branch] gh/ezyang/3133/orig -> origin/gh/ezyang/3133/orig 2025-09-07T07:28:18.7196195Z * [new branch] gh/ezyang/3134/base -> origin/gh/ezyang/3134/base 2025-09-07T07:28:18.7197062Z * [new branch] gh/ezyang/3134/head -> origin/gh/ezyang/3134/head 2025-09-07T07:28:18.7197900Z * [new branch] gh/ezyang/3134/orig -> origin/gh/ezyang/3134/orig 2025-09-07T07:28:18.7199155Z * [new branch] gh/ezyang/3135/base -> origin/gh/ezyang/3135/base 2025-09-07T07:28:18.7199946Z * [new branch] gh/ezyang/3135/head -> origin/gh/ezyang/3135/head 2025-09-07T07:28:18.7200875Z * [new branch] gh/ezyang/3135/orig -> origin/gh/ezyang/3135/orig 2025-09-07T07:28:18.7202079Z * [new branch] gh/ezyang/3136/base -> origin/gh/ezyang/3136/base 2025-09-07T07:28:18.7203021Z * [new branch] gh/ezyang/3136/head -> origin/gh/ezyang/3136/head 2025-09-07T07:28:18.7203768Z * [new branch] gh/ezyang/3136/orig -> origin/gh/ezyang/3136/orig 2025-09-07T07:28:18.7204975Z * [new branch] gh/ezyang/3137/base -> origin/gh/ezyang/3137/base 2025-09-07T07:28:18.7205838Z * [new branch] gh/ezyang/3137/head -> origin/gh/ezyang/3137/head 2025-09-07T07:28:18.7206644Z * [new branch] gh/ezyang/3137/orig -> origin/gh/ezyang/3137/orig 2025-09-07T07:28:18.7207831Z * [new branch] gh/ezyang/3138/base -> origin/gh/ezyang/3138/base 2025-09-07T07:28:18.7208928Z * [new branch] gh/ezyang/3138/head -> origin/gh/ezyang/3138/head 2025-09-07T07:28:18.7209904Z * [new branch] gh/ezyang/3138/orig -> origin/gh/ezyang/3138/orig 2025-09-07T07:28:18.7211086Z * [new branch] gh/ezyang/3139/base -> origin/gh/ezyang/3139/base 2025-09-07T07:28:18.7211933Z * [new branch] gh/ezyang/3139/head -> origin/gh/ezyang/3139/head 2025-09-07T07:28:18.7212791Z * [new branch] gh/ezyang/3139/orig -> origin/gh/ezyang/3139/orig 2025-09-07T07:28:18.7214001Z * [new branch] gh/ezyang/3140/base -> origin/gh/ezyang/3140/base 2025-09-07T07:28:18.7215140Z * [new branch] gh/ezyang/3140/head -> origin/gh/ezyang/3140/head 2025-09-07T07:28:18.7216002Z * [new branch] gh/ezyang/3140/orig -> origin/gh/ezyang/3140/orig 2025-09-07T07:28:18.7217266Z * [new branch] gh/ezyang/3141/base -> origin/gh/ezyang/3141/base 2025-09-07T07:28:18.7218215Z * [new branch] gh/ezyang/3141/head -> origin/gh/ezyang/3141/head 2025-09-07T07:28:18.7219064Z * [new branch] gh/ezyang/3141/orig -> origin/gh/ezyang/3141/orig 2025-09-07T07:28:18.7220269Z * [new branch] gh/ezyang/3142/base -> origin/gh/ezyang/3142/base 2025-09-07T07:28:18.7221085Z * [new branch] gh/ezyang/3142/head -> origin/gh/ezyang/3142/head 2025-09-07T07:28:18.7222162Z * [new branch] gh/ezyang/3142/orig -> origin/gh/ezyang/3142/orig 2025-09-07T07:28:18.7223335Z * [new branch] gh/ezyang/3143/base -> origin/gh/ezyang/3143/base 2025-09-07T07:28:18.7224245Z * [new branch] gh/ezyang/3143/head -> origin/gh/ezyang/3143/head 2025-09-07T07:28:18.7225133Z * [new branch] gh/ezyang/3143/orig -> origin/gh/ezyang/3143/orig 2025-09-07T07:28:18.7226592Z * [new branch] gh/fadara01/1/base -> origin/gh/fadara01/1/base 2025-09-07T07:28:18.7228428Z * [new branch] gh/fadara01/1/head -> origin/gh/fadara01/1/head 2025-09-07T07:28:18.7229307Z * [new branch] gh/fadara01/1/orig -> origin/gh/fadara01/1/orig 2025-09-07T07:28:18.7231054Z * [new branch] gh/fduwjj/171/base -> origin/gh/fduwjj/171/base 2025-09-07T07:28:18.7232001Z * [new branch] gh/fduwjj/171/head -> origin/gh/fduwjj/171/head 2025-09-07T07:28:18.7232843Z * [new branch] gh/fduwjj/171/orig -> origin/gh/fduwjj/171/orig 2025-09-07T07:28:18.7234169Z * [new branch] gh/fduwjj/175/base -> origin/gh/fduwjj/175/base 2025-09-07T07:28:18.7235192Z * [new branch] gh/fduwjj/175/head -> origin/gh/fduwjj/175/head 2025-09-07T07:28:18.7236046Z * [new branch] gh/fduwjj/175/orig -> origin/gh/fduwjj/175/orig 2025-09-07T07:28:18.7237298Z * [new branch] gh/fduwjj/176/base -> origin/gh/fduwjj/176/base 2025-09-07T07:28:18.7238071Z * [new branch] gh/fduwjj/176/head -> origin/gh/fduwjj/176/head 2025-09-07T07:28:18.7238928Z * [new branch] gh/fduwjj/176/orig -> origin/gh/fduwjj/176/orig 2025-09-07T07:28:18.7240075Z * [new branch] gh/fduwjj/177/base -> origin/gh/fduwjj/177/base 2025-09-07T07:28:18.7241068Z * [new branch] gh/fduwjj/177/head -> origin/gh/fduwjj/177/head 2025-09-07T07:28:18.7241794Z * [new branch] gh/fduwjj/177/orig -> origin/gh/fduwjj/177/orig 2025-09-07T07:28:18.7243052Z * [new branch] gh/fduwjj/178/base -> origin/gh/fduwjj/178/base 2025-09-07T07:28:18.7244031Z * [new branch] gh/fduwjj/178/head -> origin/gh/fduwjj/178/head 2025-09-07T07:28:18.7244876Z * [new branch] gh/fduwjj/178/orig -> origin/gh/fduwjj/178/orig 2025-09-07T07:28:18.7246035Z * [new branch] gh/fduwjj/179/base -> origin/gh/fduwjj/179/base 2025-09-07T07:28:18.7246871Z * [new branch] gh/fduwjj/179/head -> origin/gh/fduwjj/179/head 2025-09-07T07:28:18.7247744Z * [new branch] gh/fduwjj/179/orig -> origin/gh/fduwjj/179/orig 2025-09-07T07:28:18.7248969Z * [new branch] gh/fduwjj/180/base -> origin/gh/fduwjj/180/base 2025-09-07T07:28:18.7249784Z * [new branch] gh/fduwjj/180/head -> origin/gh/fduwjj/180/head 2025-09-07T07:28:18.7250620Z * [new branch] gh/fduwjj/180/orig -> origin/gh/fduwjj/180/orig 2025-09-07T07:28:18.7251804Z * [new branch] gh/fduwjj/181/base -> origin/gh/fduwjj/181/base 2025-09-07T07:28:18.7252708Z * [new branch] gh/fduwjj/181/head -> origin/gh/fduwjj/181/head 2025-09-07T07:28:18.7253565Z * [new branch] gh/fduwjj/181/orig -> origin/gh/fduwjj/181/orig 2025-09-07T07:28:18.7254868Z * [new branch] gh/fduwjj/182/base -> origin/gh/fduwjj/182/base 2025-09-07T07:28:18.7255698Z * [new branch] gh/fduwjj/182/head -> origin/gh/fduwjj/182/head 2025-09-07T07:28:18.7256513Z * [new branch] gh/fduwjj/182/orig -> origin/gh/fduwjj/182/orig 2025-09-07T07:28:18.7257790Z * [new branch] gh/fduwjj/183/base -> origin/gh/fduwjj/183/base 2025-09-07T07:28:18.7258829Z * [new branch] gh/fduwjj/183/head -> origin/gh/fduwjj/183/head 2025-09-07T07:28:18.7259687Z * [new branch] gh/fduwjj/183/orig -> origin/gh/fduwjj/183/orig 2025-09-07T07:28:18.7261097Z * [new branch] gh/fduwjj/184/base -> origin/gh/fduwjj/184/base 2025-09-07T07:28:18.7261950Z * [new branch] gh/fduwjj/184/head -> origin/gh/fduwjj/184/head 2025-09-07T07:28:18.7262753Z * [new branch] gh/fduwjj/184/orig -> origin/gh/fduwjj/184/orig 2025-09-07T07:28:18.7264038Z * [new branch] gh/fduwjj/185/base -> origin/gh/fduwjj/185/base 2025-09-07T07:28:18.7264915Z * [new branch] gh/fduwjj/185/head -> origin/gh/fduwjj/185/head 2025-09-07T07:28:18.7265882Z * [new branch] gh/fduwjj/185/orig -> origin/gh/fduwjj/185/orig 2025-09-07T07:28:18.7267048Z * [new branch] gh/fduwjj/186/base -> origin/gh/fduwjj/186/base 2025-09-07T07:28:18.7267910Z * [new branch] gh/fduwjj/186/head -> origin/gh/fduwjj/186/head 2025-09-07T07:28:18.7268791Z * [new branch] gh/fduwjj/186/orig -> origin/gh/fduwjj/186/orig 2025-09-07T07:28:18.7269809Z * [new branch] gh/fduwjj/187/base -> origin/gh/fduwjj/187/base 2025-09-07T07:28:18.7270671Z * [new branch] gh/fduwjj/187/head -> origin/gh/fduwjj/187/head 2025-09-07T07:28:18.7271526Z * [new branch] gh/fduwjj/187/orig -> origin/gh/fduwjj/187/orig 2025-09-07T07:28:18.7272627Z * [new branch] gh/fduwjj/188/base -> origin/gh/fduwjj/188/base 2025-09-07T07:28:18.7273420Z * [new branch] gh/fduwjj/188/head -> origin/gh/fduwjj/188/head 2025-09-07T07:28:18.7274194Z * [new branch] gh/fduwjj/188/orig -> origin/gh/fduwjj/188/orig 2025-09-07T07:28:18.7275394Z * [new branch] gh/fduwjj/189/base -> origin/gh/fduwjj/189/base 2025-09-07T07:28:18.7275997Z * [new branch] gh/fduwjj/189/head -> origin/gh/fduwjj/189/head 2025-09-07T07:28:18.7276780Z * [new branch] gh/fduwjj/189/orig -> origin/gh/fduwjj/189/orig 2025-09-07T07:28:18.7278281Z * [new branch] gh/fduwjj/190/base -> origin/gh/fduwjj/190/base 2025-09-07T07:28:18.7279188Z * [new branch] gh/fduwjj/190/head -> origin/gh/fduwjj/190/head 2025-09-07T07:28:18.7280067Z * [new branch] gh/fduwjj/190/orig -> origin/gh/fduwjj/190/orig 2025-09-07T07:28:18.7281180Z * [new branch] gh/fduwjj/191/base -> origin/gh/fduwjj/191/base 2025-09-07T07:28:18.7282168Z * [new branch] gh/fduwjj/191/head -> origin/gh/fduwjj/191/head 2025-09-07T07:28:18.7283042Z * [new branch] gh/fduwjj/191/orig -> origin/gh/fduwjj/191/orig 2025-09-07T07:28:18.7284441Z * [new branch] gh/fegin/306/base -> origin/gh/fegin/306/base 2025-09-07T07:28:18.7285346Z * [new branch] gh/fegin/306/head -> origin/gh/fegin/306/head 2025-09-07T07:28:18.7286326Z * [new branch] gh/fegin/306/orig -> origin/gh/fegin/306/orig 2025-09-07T07:28:18.7287363Z * [new branch] gh/fegin/307/base -> origin/gh/fegin/307/base 2025-09-07T07:28:18.7288370Z * [new branch] gh/fegin/307/head -> origin/gh/fegin/307/head 2025-09-07T07:28:18.7289200Z * [new branch] gh/fegin/307/orig -> origin/gh/fegin/307/orig 2025-09-07T07:28:18.7290367Z * [new branch] gh/fegin/308/base -> origin/gh/fegin/308/base 2025-09-07T07:28:18.7291259Z * [new branch] gh/fegin/308/head -> origin/gh/fegin/308/head 2025-09-07T07:28:18.7292150Z * [new branch] gh/fegin/308/orig -> origin/gh/fegin/308/orig 2025-09-07T07:28:18.7293391Z * [new branch] gh/fegin/309/base -> origin/gh/fegin/309/base 2025-09-07T07:28:18.7294188Z * [new branch] gh/fegin/309/head -> origin/gh/fegin/309/head 2025-09-07T07:28:18.7295118Z * [new branch] gh/fegin/309/orig -> origin/gh/fegin/309/orig 2025-09-07T07:28:18.7296281Z * [new branch] gh/fegin/310/base -> origin/gh/fegin/310/base 2025-09-07T07:28:18.7297120Z * [new branch] gh/fegin/310/head -> origin/gh/fegin/310/head 2025-09-07T07:28:18.7298043Z * [new branch] gh/fegin/310/orig -> origin/gh/fegin/310/orig 2025-09-07T07:28:18.7299152Z * [new branch] gh/fegin/311/base -> origin/gh/fegin/311/base 2025-09-07T07:28:18.7300030Z * [new branch] gh/fegin/311/head -> origin/gh/fegin/311/head 2025-09-07T07:28:18.7300981Z * [new branch] gh/fegin/311/orig -> origin/gh/fegin/311/orig 2025-09-07T07:28:18.7303082Z * [new branch] gh/fegin/312/base -> origin/gh/fegin/312/base 2025-09-07T07:28:18.7303260Z * [new branch] gh/fegin/312/head -> origin/gh/fegin/312/head 2025-09-07T07:28:18.7303747Z * [new branch] gh/fegin/312/orig -> origin/gh/fegin/312/orig 2025-09-07T07:28:18.7305149Z * [new branch] gh/fegin/313/base -> origin/gh/fegin/313/base 2025-09-07T07:28:18.7305862Z * [new branch] gh/fegin/313/head -> origin/gh/fegin/313/head 2025-09-07T07:28:18.7306725Z * [new branch] gh/fegin/313/orig -> origin/gh/fegin/313/orig 2025-09-07T07:28:18.7308194Z * [new branch] gh/fffrog/124/base -> origin/gh/fffrog/124/base 2025-09-07T07:28:18.7310323Z * [new branch] gh/fffrog/124/head -> origin/gh/fffrog/124/head 2025-09-07T07:28:18.7310512Z * [new branch] gh/fffrog/124/orig -> origin/gh/fffrog/124/orig 2025-09-07T07:28:18.7311557Z * [new branch] gh/fffrog/129/base -> origin/gh/fffrog/129/base 2025-09-07T07:28:18.7312116Z * [new branch] gh/fffrog/129/head -> origin/gh/fffrog/129/head 2025-09-07T07:28:18.7312922Z * [new branch] gh/fffrog/129/orig -> origin/gh/fffrog/129/orig 2025-09-07T07:28:18.7314140Z * [new branch] gh/fffrog/130/base -> origin/gh/fffrog/130/base 2025-09-07T07:28:18.7315028Z * [new branch] gh/fffrog/130/head -> origin/gh/fffrog/130/head 2025-09-07T07:28:18.7315961Z * [new branch] gh/fffrog/130/orig -> origin/gh/fffrog/130/orig 2025-09-07T07:28:18.7317078Z * [new branch] gh/fffrog/131/base -> origin/gh/fffrog/131/base 2025-09-07T07:28:18.7317976Z * [new branch] gh/fffrog/131/head -> origin/gh/fffrog/131/head 2025-09-07T07:28:18.7318846Z * [new branch] gh/fffrog/131/orig -> origin/gh/fffrog/131/orig 2025-09-07T07:28:18.7319985Z * [new branch] gh/fffrog/132/base -> origin/gh/fffrog/132/base 2025-09-07T07:28:18.7320865Z * [new branch] gh/fffrog/132/head -> origin/gh/fffrog/132/head 2025-09-07T07:28:18.7321721Z * [new branch] gh/fffrog/132/orig -> origin/gh/fffrog/132/orig 2025-09-07T07:28:18.7322833Z * [new branch] gh/fffrog/133/base -> origin/gh/fffrog/133/base 2025-09-07T07:28:18.7323772Z * [new branch] gh/fffrog/133/head -> origin/gh/fffrog/133/head 2025-09-07T07:28:18.7324638Z * [new branch] gh/fffrog/133/orig -> origin/gh/fffrog/133/orig 2025-09-07T07:28:18.7325792Z * [new branch] gh/fffrog/134/base -> origin/gh/fffrog/134/base 2025-09-07T07:28:18.7326671Z * [new branch] gh/fffrog/134/head -> origin/gh/fffrog/134/head 2025-09-07T07:28:18.7327519Z * [new branch] gh/fffrog/134/orig -> origin/gh/fffrog/134/orig 2025-09-07T07:28:18.7328717Z * [new branch] gh/fffrog/135/base -> origin/gh/fffrog/135/base 2025-09-07T07:28:18.7329502Z * [new branch] gh/fffrog/135/head -> origin/gh/fffrog/135/head 2025-09-07T07:28:18.7330355Z * [new branch] gh/fffrog/135/orig -> origin/gh/fffrog/135/orig 2025-09-07T07:28:18.7331525Z * [new branch] gh/fffrog/136/base -> origin/gh/fffrog/136/base 2025-09-07T07:28:18.7332376Z * [new branch] gh/fffrog/136/head -> origin/gh/fffrog/136/head 2025-09-07T07:28:18.7333199Z * [new branch] gh/fffrog/136/orig -> origin/gh/fffrog/136/orig 2025-09-07T07:28:18.7334295Z * [new branch] gh/fffrog/137/base -> origin/gh/fffrog/137/base 2025-09-07T07:28:18.7335241Z * [new branch] gh/fffrog/137/head -> origin/gh/fffrog/137/head 2025-09-07T07:28:18.7336092Z * [new branch] gh/fffrog/137/orig -> origin/gh/fffrog/137/orig 2025-09-07T07:28:18.7337280Z * [new branch] gh/fffrog/138/base -> origin/gh/fffrog/138/base 2025-09-07T07:28:18.7338075Z * [new branch] gh/fffrog/138/head -> origin/gh/fffrog/138/head 2025-09-07T07:28:18.7338944Z * [new branch] gh/fffrog/138/orig -> origin/gh/fffrog/138/orig 2025-09-07T07:28:18.7340098Z * [new branch] gh/fffrog/139/base -> origin/gh/fffrog/139/base 2025-09-07T07:28:18.7340993Z * [new branch] gh/fffrog/139/head -> origin/gh/fffrog/139/head 2025-09-07T07:28:18.7341879Z * [new branch] gh/fffrog/139/orig -> origin/gh/fffrog/139/orig 2025-09-07T07:28:18.7343355Z * [new branch] gh/fffrog/140/base -> origin/gh/fffrog/140/base 2025-09-07T07:28:18.7343997Z * [new branch] gh/fffrog/140/head -> origin/gh/fffrog/140/head 2025-09-07T07:28:18.7344825Z * [new branch] gh/fffrog/140/orig -> origin/gh/fffrog/140/orig 2025-09-07T07:28:18.7346046Z * [new branch] gh/fffrog/141/base -> origin/gh/fffrog/141/base 2025-09-07T07:28:18.7346860Z * [new branch] gh/fffrog/141/head -> origin/gh/fffrog/141/head 2025-09-07T07:28:18.7347711Z * [new branch] gh/fffrog/141/orig -> origin/gh/fffrog/141/orig 2025-09-07T07:28:18.7348957Z * [new branch] gh/fffrog/142/base -> origin/gh/fffrog/142/base 2025-09-07T07:28:18.7349812Z * [new branch] gh/fffrog/142/head -> origin/gh/fffrog/142/head 2025-09-07T07:28:18.7350660Z * [new branch] gh/fffrog/142/orig -> origin/gh/fffrog/142/orig 2025-09-07T07:28:18.7351793Z * [new branch] gh/fffrog/143/base -> origin/gh/fffrog/143/base 2025-09-07T07:28:18.7352666Z * [new branch] gh/fffrog/143/head -> origin/gh/fffrog/143/head 2025-09-07T07:28:18.7353524Z * [new branch] gh/fffrog/143/orig -> origin/gh/fffrog/143/orig 2025-09-07T07:28:18.7355004Z * [new branch] gh/fffrog/144/base -> origin/gh/fffrog/144/base 2025-09-07T07:28:18.7355950Z * [new branch] gh/fffrog/144/head -> origin/gh/fffrog/144/head 2025-09-07T07:28:18.7356813Z * [new branch] gh/fffrog/144/orig -> origin/gh/fffrog/144/orig 2025-09-07T07:28:18.7357999Z * [new branch] gh/fffrog/145/base -> origin/gh/fffrog/145/base 2025-09-07T07:28:18.7358936Z * [new branch] gh/fffrog/145/head -> origin/gh/fffrog/145/head 2025-09-07T07:28:18.7359784Z * [new branch] gh/fffrog/145/orig -> origin/gh/fffrog/145/orig 2025-09-07T07:28:18.7360966Z * [new branch] gh/fffrog/146/base -> origin/gh/fffrog/146/base 2025-09-07T07:28:18.7361817Z * [new branch] gh/fffrog/146/head -> origin/gh/fffrog/146/head 2025-09-07T07:28:18.7362661Z * [new branch] gh/fffrog/146/orig -> origin/gh/fffrog/146/orig 2025-09-07T07:28:18.7363887Z * [new branch] gh/fffrog/147/base -> origin/gh/fffrog/147/base 2025-09-07T07:28:18.7364722Z * [new branch] gh/fffrog/147/head -> origin/gh/fffrog/147/head 2025-09-07T07:28:18.7365666Z * [new branch] gh/fffrog/147/orig -> origin/gh/fffrog/147/orig 2025-09-07T07:28:18.7366915Z * [new branch] gh/fffrog/148/base -> origin/gh/fffrog/148/base 2025-09-07T07:28:18.7367757Z * [new branch] gh/fffrog/148/head -> origin/gh/fffrog/148/head 2025-09-07T07:28:18.7368628Z * [new branch] gh/fffrog/148/orig -> origin/gh/fffrog/148/orig 2025-09-07T07:28:18.7369886Z * [new branch] gh/fffrog/149/base -> origin/gh/fffrog/149/base 2025-09-07T07:28:18.7370833Z * [new branch] gh/fffrog/149/head -> origin/gh/fffrog/149/head 2025-09-07T07:28:18.7371787Z * [new branch] gh/fffrog/149/orig -> origin/gh/fffrog/149/orig 2025-09-07T07:28:18.7372938Z * [new branch] gh/fffrog/150/base -> origin/gh/fffrog/150/base 2025-09-07T07:28:18.7373785Z * [new branch] gh/fffrog/150/head -> origin/gh/fffrog/150/head 2025-09-07T07:28:18.7374825Z * [new branch] gh/fffrog/150/orig -> origin/gh/fffrog/150/orig 2025-09-07T07:28:18.7375873Z * [new branch] gh/fffrog/151/base -> origin/gh/fffrog/151/base 2025-09-07T07:28:18.7376704Z * [new branch] gh/fffrog/151/head -> origin/gh/fffrog/151/head 2025-09-07T07:28:18.7377602Z * [new branch] gh/fffrog/151/orig -> origin/gh/fffrog/151/orig 2025-09-07T07:28:18.7378765Z * [new branch] gh/fffrog/152/base -> origin/gh/fffrog/152/base 2025-09-07T07:28:18.7379613Z * [new branch] gh/fffrog/152/head -> origin/gh/fffrog/152/head 2025-09-07T07:28:18.7380913Z * [new branch] gh/fffrog/153/base -> origin/gh/fffrog/153/base 2025-09-07T07:28:18.7381660Z * [new branch] gh/fffrog/153/head -> origin/gh/fffrog/153/head 2025-09-07T07:28:18.7382911Z * [new branch] gh/fffrog/153/orig -> origin/gh/fffrog/153/orig 2025-09-07T07:28:18.7384079Z * [new branch] gh/gmagogsfm/1/base -> origin/gh/gmagogsfm/1/base 2025-09-07T07:28:18.7384923Z * [new branch] gh/gmagogsfm/1/head -> origin/gh/gmagogsfm/1/head 2025-09-07T07:28:18.7385788Z * [new branch] gh/gmagogsfm/1/orig -> origin/gh/gmagogsfm/1/orig 2025-09-07T07:28:18.7386909Z * [new branch] gh/gmagogsfm/2/base -> origin/gh/gmagogsfm/2/base 2025-09-07T07:28:18.7387719Z * [new branch] gh/gmagogsfm/2/head -> origin/gh/gmagogsfm/2/head 2025-09-07T07:28:18.7388524Z * [new branch] gh/gmagogsfm/2/orig -> origin/gh/gmagogsfm/2/orig 2025-09-07T07:28:18.7389640Z * [new branch] gh/gmagogsfm/3/base -> origin/gh/gmagogsfm/3/base 2025-09-07T07:28:18.7390500Z * [new branch] gh/gmagogsfm/3/head -> origin/gh/gmagogsfm/3/head 2025-09-07T07:28:18.7391323Z * [new branch] gh/gmagogsfm/3/orig -> origin/gh/gmagogsfm/3/orig 2025-09-07T07:28:18.7392896Z * [new branch] gh/guangyey/134/base -> origin/gh/guangyey/134/base 2025-09-07T07:28:18.7393763Z * [new branch] gh/guangyey/134/head -> origin/gh/guangyey/134/head 2025-09-07T07:28:18.7394573Z * [new branch] gh/guangyey/134/orig -> origin/gh/guangyey/134/orig 2025-09-07T07:28:18.7395773Z * [new branch] gh/guangyey/135/base -> origin/gh/guangyey/135/base 2025-09-07T07:28:18.7396669Z * [new branch] gh/guangyey/135/head -> origin/gh/guangyey/135/head 2025-09-07T07:28:18.7397483Z * [new branch] gh/guangyey/135/orig -> origin/gh/guangyey/135/orig 2025-09-07T07:28:18.7398699Z * [new branch] gh/guangyey/139/base -> origin/gh/guangyey/139/base 2025-09-07T07:28:18.7399551Z * [new branch] gh/guangyey/139/head -> origin/gh/guangyey/139/head 2025-09-07T07:28:18.7400381Z * [new branch] gh/guangyey/139/orig -> origin/gh/guangyey/139/orig 2025-09-07T07:28:18.7401616Z * [new branch] gh/guangyey/140/base -> origin/gh/guangyey/140/base 2025-09-07T07:28:18.7402483Z * [new branch] gh/guangyey/140/head -> origin/gh/guangyey/140/head 2025-09-07T07:28:18.7403366Z * [new branch] gh/guangyey/140/orig -> origin/gh/guangyey/140/orig 2025-09-07T07:28:18.7404483Z * [new branch] gh/guangyey/142/base -> origin/gh/guangyey/142/base 2025-09-07T07:28:18.7405338Z * [new branch] gh/guangyey/142/head -> origin/gh/guangyey/142/head 2025-09-07T07:28:18.7406196Z * [new branch] gh/guangyey/142/orig -> origin/gh/guangyey/142/orig 2025-09-07T07:28:18.7407342Z * [new branch] gh/guangyey/145/base -> origin/gh/guangyey/145/base 2025-09-07T07:28:18.7408209Z * [new branch] gh/guangyey/145/head -> origin/gh/guangyey/145/head 2025-09-07T07:28:18.7409007Z * [new branch] gh/guangyey/145/orig -> origin/gh/guangyey/145/orig 2025-09-07T07:28:18.7410141Z * [new branch] gh/guangyey/153/base -> origin/gh/guangyey/153/base 2025-09-07T07:28:18.7410973Z * [new branch] gh/guangyey/153/head -> origin/gh/guangyey/153/head 2025-09-07T07:28:18.7411878Z * [new branch] gh/guangyey/153/orig -> origin/gh/guangyey/153/orig 2025-09-07T07:28:18.7413019Z * [new branch] gh/guangyey/159/base -> origin/gh/guangyey/159/base 2025-09-07T07:28:18.7413864Z * [new branch] gh/guangyey/159/head -> origin/gh/guangyey/159/head 2025-09-07T07:28:18.7414876Z * [new branch] gh/guangyey/159/orig -> origin/gh/guangyey/159/orig 2025-09-07T07:28:18.7415967Z * [new branch] gh/guangyey/163/base -> origin/gh/guangyey/163/base 2025-09-07T07:28:18.7416836Z * [new branch] gh/guangyey/163/head -> origin/gh/guangyey/163/head 2025-09-07T07:28:18.7417886Z * [new branch] gh/guangyey/163/orig -> origin/gh/guangyey/163/orig 2025-09-07T07:28:18.7419109Z * [new branch] gh/guangyey/168/base -> origin/gh/guangyey/168/base 2025-09-07T07:28:18.7419962Z * [new branch] gh/guangyey/168/head -> origin/gh/guangyey/168/head 2025-09-07T07:28:18.7420791Z * [new branch] gh/guangyey/168/orig -> origin/gh/guangyey/168/orig 2025-09-07T07:28:18.7421940Z * [new branch] gh/guangyey/169/base -> origin/gh/guangyey/169/base 2025-09-07T07:28:18.7422772Z * [new branch] gh/guangyey/169/head -> origin/gh/guangyey/169/head 2025-09-07T07:28:18.7423615Z * [new branch] gh/guangyey/169/orig -> origin/gh/guangyey/169/orig 2025-09-07T07:28:18.7424851Z * [new branch] gh/guangyey/170/base -> origin/gh/guangyey/170/base 2025-09-07T07:28:18.7425714Z * [new branch] gh/guangyey/170/head -> origin/gh/guangyey/170/head 2025-09-07T07:28:18.7426545Z * [new branch] gh/guangyey/170/orig -> origin/gh/guangyey/170/orig 2025-09-07T07:28:18.7427789Z * [new branch] gh/guangyey/171/base -> origin/gh/guangyey/171/base 2025-09-07T07:28:18.7428676Z * [new branch] gh/guangyey/171/head -> origin/gh/guangyey/171/head 2025-09-07T07:28:18.7429502Z * [new branch] gh/guangyey/171/orig -> origin/gh/guangyey/171/orig 2025-09-07T07:28:18.7430755Z * [new branch] gh/guangyey/174/base -> origin/gh/guangyey/174/base 2025-09-07T07:28:18.7431546Z * [new branch] gh/guangyey/174/head -> origin/gh/guangyey/174/head 2025-09-07T07:28:18.7432468Z * [new branch] gh/guangyey/174/orig -> origin/gh/guangyey/174/orig 2025-09-07T07:28:18.7433607Z * [new branch] gh/guangyey/176/base -> origin/gh/guangyey/176/base 2025-09-07T07:28:18.7434524Z * [new branch] gh/guangyey/176/head -> origin/gh/guangyey/176/head 2025-09-07T07:28:18.7435367Z * [new branch] gh/guangyey/176/orig -> origin/gh/guangyey/176/orig 2025-09-07T07:28:18.7436522Z * [new branch] gh/guangyey/178/base -> origin/gh/guangyey/178/base 2025-09-07T07:28:18.7437297Z * [new branch] gh/guangyey/178/head -> origin/gh/guangyey/178/head 2025-09-07T07:28:18.7438189Z * [new branch] gh/guangyey/178/orig -> origin/gh/guangyey/178/orig 2025-09-07T07:28:18.7439793Z * [new branch] gh/guangyey/181/base -> origin/gh/guangyey/181/base 2025-09-07T07:28:18.7440679Z * [new branch] gh/guangyey/181/head -> origin/gh/guangyey/181/head 2025-09-07T07:28:18.7441519Z * [new branch] gh/guangyey/181/orig -> origin/gh/guangyey/181/orig 2025-09-07T07:28:18.7442701Z * [new branch] gh/guangyey/182/base -> origin/gh/guangyey/182/base 2025-09-07T07:28:18.7443644Z * [new branch] gh/guangyey/182/head -> origin/gh/guangyey/182/head 2025-09-07T07:28:18.7444490Z * [new branch] gh/guangyey/182/orig -> origin/gh/guangyey/182/orig 2025-09-07T07:28:18.7445615Z * [new branch] gh/guangyey/183/base -> origin/gh/guangyey/183/base 2025-09-07T07:28:18.7446494Z * [new branch] gh/guangyey/183/head -> origin/gh/guangyey/183/head 2025-09-07T07:28:18.7447384Z * [new branch] gh/guangyey/183/orig -> origin/gh/guangyey/183/orig 2025-09-07T07:28:18.7448568Z * [new branch] gh/guangyey/184/base -> origin/gh/guangyey/184/base 2025-09-07T07:28:18.7449486Z * [new branch] gh/guangyey/184/head -> origin/gh/guangyey/184/head 2025-09-07T07:28:18.7450232Z * [new branch] gh/guangyey/184/orig -> origin/gh/guangyey/184/orig 2025-09-07T07:28:18.7451520Z * [new branch] gh/guangyey/185/base -> origin/gh/guangyey/185/base 2025-09-07T07:28:18.7452391Z * [new branch] gh/guangyey/185/head -> origin/gh/guangyey/185/head 2025-09-07T07:28:18.7453246Z * [new branch] gh/guangyey/185/orig -> origin/gh/guangyey/185/orig 2025-09-07T07:28:18.7454414Z * [new branch] gh/guangyey/186/base -> origin/gh/guangyey/186/base 2025-09-07T07:28:18.7455560Z * [new branch] gh/guangyey/186/head -> origin/gh/guangyey/186/head 2025-09-07T07:28:18.7456341Z * [new branch] gh/guangyey/186/orig -> origin/gh/guangyey/186/orig 2025-09-07T07:28:18.7457600Z * [new branch] gh/guangyey/187/base -> origin/gh/guangyey/187/base 2025-09-07T07:28:18.7458449Z * [new branch] gh/guangyey/187/head -> origin/gh/guangyey/187/head 2025-09-07T07:28:18.7459275Z * [new branch] gh/guangyey/187/orig -> origin/gh/guangyey/187/orig 2025-09-07T07:28:18.7460580Z * [new branch] gh/guangyey/188/base -> origin/gh/guangyey/188/base 2025-09-07T07:28:18.7461681Z * [new branch] gh/guangyey/188/head -> origin/gh/guangyey/188/head 2025-09-07T07:28:18.7462576Z * [new branch] gh/guangyey/188/orig -> origin/gh/guangyey/188/orig 2025-09-07T07:28:18.7463836Z * [new branch] gh/guangyey/189/base -> origin/gh/guangyey/189/base 2025-09-07T07:28:18.7464704Z * [new branch] gh/guangyey/189/head -> origin/gh/guangyey/189/head 2025-09-07T07:28:18.7465575Z * [new branch] gh/guangyey/189/orig -> origin/gh/guangyey/189/orig 2025-09-07T07:28:18.7466820Z * [new branch] gh/guangyey/190/base -> origin/gh/guangyey/190/base 2025-09-07T07:28:18.7467658Z * [new branch] gh/guangyey/190/head -> origin/gh/guangyey/190/head 2025-09-07T07:28:18.7468541Z * [new branch] gh/guangyey/190/orig -> origin/gh/guangyey/190/orig 2025-09-07T07:28:18.7469737Z * [new branch] gh/guangyey/191/base -> origin/gh/guangyey/191/base 2025-09-07T07:28:18.7470590Z * [new branch] gh/guangyey/191/head -> origin/gh/guangyey/191/head 2025-09-07T07:28:18.7471795Z * [new branch] gh/guangyey/191/orig -> origin/gh/guangyey/191/orig 2025-09-07T07:28:18.7473030Z * [new branch] gh/guangyey/192/base -> origin/gh/guangyey/192/base 2025-09-07T07:28:18.7473898Z * [new branch] gh/guangyey/192/head -> origin/gh/guangyey/192/head 2025-09-07T07:28:18.7474818Z * [new branch] gh/guangyey/192/orig -> origin/gh/guangyey/192/orig 2025-09-07T07:28:18.7476054Z * [new branch] gh/guangyey/193/base -> origin/gh/guangyey/193/base 2025-09-07T07:28:18.7476941Z * [new branch] gh/guangyey/193/head -> origin/gh/guangyey/193/head 2025-09-07T07:28:18.7477813Z * [new branch] gh/guangyey/193/orig -> origin/gh/guangyey/193/orig 2025-09-07T07:28:18.7479001Z * [new branch] gh/guangyey/194/base -> origin/gh/guangyey/194/base 2025-09-07T07:28:18.7479832Z * [new branch] gh/guangyey/194/head -> origin/gh/guangyey/194/head 2025-09-07T07:28:18.7480760Z * [new branch] gh/guangyey/194/orig -> origin/gh/guangyey/194/orig 2025-09-07T07:28:18.7481937Z * [new branch] gh/guangyey/195/base -> origin/gh/guangyey/195/base 2025-09-07T07:28:18.7482875Z * [new branch] gh/guangyey/195/head -> origin/gh/guangyey/195/head 2025-09-07T07:28:18.7483753Z * [new branch] gh/guangyey/195/orig -> origin/gh/guangyey/195/orig 2025-09-07T07:28:18.7485039Z * [new branch] gh/guangyey/196/base -> origin/gh/guangyey/196/base 2025-09-07T07:28:18.7485807Z * [new branch] gh/guangyey/196/head -> origin/gh/guangyey/196/head 2025-09-07T07:28:18.7486689Z * [new branch] gh/guangyey/196/orig -> origin/gh/guangyey/196/orig 2025-09-07T07:28:18.7487965Z * [new branch] gh/guangyey/197/base -> origin/gh/guangyey/197/base 2025-09-07T07:28:18.7488882Z * [new branch] gh/guangyey/197/head -> origin/gh/guangyey/197/head 2025-09-07T07:28:18.7489821Z * [new branch] gh/guangyey/197/orig -> origin/gh/guangyey/197/orig 2025-09-07T07:28:18.7491020Z * [new branch] gh/guangyey/198/base -> origin/gh/guangyey/198/base 2025-09-07T07:28:18.7491938Z * [new branch] gh/guangyey/198/head -> origin/gh/guangyey/198/head 2025-09-07T07:28:18.7493016Z * [new branch] gh/guangyey/198/orig -> origin/gh/guangyey/198/orig 2025-09-07T07:28:18.7494261Z * [new branch] gh/guangyey/199/base -> origin/gh/guangyey/199/base 2025-09-07T07:28:18.7495238Z * [new branch] gh/guangyey/199/head -> origin/gh/guangyey/199/head 2025-09-07T07:28:18.7496152Z * [new branch] gh/guangyey/199/orig -> origin/gh/guangyey/199/orig 2025-09-07T07:28:18.7497379Z * [new branch] gh/guangyey/200/base -> origin/gh/guangyey/200/base 2025-09-07T07:28:18.7498336Z * [new branch] gh/guangyey/200/head -> origin/gh/guangyey/200/head 2025-09-07T07:28:18.7499083Z * [new branch] gh/guangyey/200/orig -> origin/gh/guangyey/200/orig 2025-09-07T07:28:18.7500304Z * [new branch] gh/guangyey/201/base -> origin/gh/guangyey/201/base 2025-09-07T07:28:18.7501234Z * [new branch] gh/guangyey/201/head -> origin/gh/guangyey/201/head 2025-09-07T07:28:18.7502106Z * [new branch] gh/guangyey/201/orig -> origin/gh/guangyey/201/orig 2025-09-07T07:28:18.7503323Z * [new branch] gh/guangyey/202/base -> origin/gh/guangyey/202/base 2025-09-07T07:28:18.7504167Z * [new branch] gh/guangyey/202/head -> origin/gh/guangyey/202/head 2025-09-07T07:28:18.7505008Z * [new branch] gh/guangyey/202/orig -> origin/gh/guangyey/202/orig 2025-09-07T07:28:18.7506168Z * [new branch] gh/guangyey/203/base -> origin/gh/guangyey/203/base 2025-09-07T07:28:18.7507030Z * [new branch] gh/guangyey/203/head -> origin/gh/guangyey/203/head 2025-09-07T07:28:18.7507883Z * [new branch] gh/guangyey/203/orig -> origin/gh/guangyey/203/orig 2025-09-07T07:28:18.7509069Z * [new branch] gh/guangyey/204/base -> origin/gh/guangyey/204/base 2025-09-07T07:28:18.7509936Z * [new branch] gh/guangyey/204/head -> origin/gh/guangyey/204/head 2025-09-07T07:28:18.7510787Z * [new branch] gh/guangyey/204/orig -> origin/gh/guangyey/204/orig 2025-09-07T07:28:18.7512028Z * [new branch] gh/guangyey/205/base -> origin/gh/guangyey/205/base 2025-09-07T07:28:18.7512836Z * [new branch] gh/guangyey/205/head -> origin/gh/guangyey/205/head 2025-09-07T07:28:18.7513710Z * [new branch] gh/guangyey/205/orig -> origin/gh/guangyey/205/orig 2025-09-07T07:28:18.7514847Z * [new branch] gh/guangyey/206/base -> origin/gh/guangyey/206/base 2025-09-07T07:28:18.7515723Z * [new branch] gh/guangyey/206/head -> origin/gh/guangyey/206/head 2025-09-07T07:28:18.7516542Z * [new branch] gh/guangyey/206/orig -> origin/gh/guangyey/206/orig 2025-09-07T07:28:18.7517731Z * [new branch] gh/guangyey/207/base -> origin/gh/guangyey/207/base 2025-09-07T07:28:18.7518552Z * [new branch] gh/guangyey/207/head -> origin/gh/guangyey/207/head 2025-09-07T07:28:18.7519612Z * [new branch] gh/guangyey/207/orig -> origin/gh/guangyey/207/orig 2025-09-07T07:28:18.7520740Z * [new branch] gh/guangyey/79/base -> origin/gh/guangyey/79/base 2025-09-07T07:28:18.7521592Z * [new branch] gh/guangyey/79/head -> origin/gh/guangyey/79/head 2025-09-07T07:28:18.7522440Z * [new branch] gh/guangyey/79/orig -> origin/gh/guangyey/79/orig 2025-09-07T07:28:18.7523718Z * [new branch] gh/guangyey/89/base -> origin/gh/guangyey/89/base 2025-09-07T07:28:18.7524570Z * [new branch] gh/guangyey/89/head -> origin/gh/guangyey/89/head 2025-09-07T07:28:18.7525431Z * [new branch] gh/guangyey/89/orig -> origin/gh/guangyey/89/orig 2025-09-07T07:28:18.7526942Z * [new branch] gh/guilhermeleobas/107/base -> origin/gh/guilhermeleobas/107/base 2025-09-07T07:28:18.7527822Z * [new branch] gh/guilhermeleobas/107/head -> origin/gh/guilhermeleobas/107/head 2025-09-07T07:28:18.7528717Z * [new branch] gh/guilhermeleobas/107/orig -> origin/gh/guilhermeleobas/107/orig 2025-09-07T07:28:18.7529879Z * [new branch] gh/guilhermeleobas/108/base -> origin/gh/guilhermeleobas/108/base 2025-09-07T07:28:18.7530703Z * [new branch] gh/guilhermeleobas/108/head -> origin/gh/guilhermeleobas/108/head 2025-09-07T07:28:18.7531542Z * [new branch] gh/guilhermeleobas/108/orig -> origin/gh/guilhermeleobas/108/orig 2025-09-07T07:28:18.7532775Z * [new branch] gh/guilhermeleobas/124/base -> origin/gh/guilhermeleobas/124/base 2025-09-07T07:28:18.7533631Z * [new branch] gh/guilhermeleobas/124/head -> origin/gh/guilhermeleobas/124/head 2025-09-07T07:28:18.7534484Z * [new branch] gh/guilhermeleobas/124/orig -> origin/gh/guilhermeleobas/124/orig 2025-09-07T07:28:18.7535847Z * [new branch] gh/guilhermeleobas/147/base -> origin/gh/guilhermeleobas/147/base 2025-09-07T07:28:18.7536575Z * [new branch] gh/guilhermeleobas/147/head -> origin/gh/guilhermeleobas/147/head 2025-09-07T07:28:18.7537441Z * [new branch] gh/guilhermeleobas/147/orig -> origin/gh/guilhermeleobas/147/orig 2025-09-07T07:28:18.7538637Z * [new branch] gh/guilhermeleobas/150/base -> origin/gh/guilhermeleobas/150/base 2025-09-07T07:28:18.7539460Z * [new branch] gh/guilhermeleobas/150/head -> origin/gh/guilhermeleobas/150/head 2025-09-07T07:28:18.7540328Z * [new branch] gh/guilhermeleobas/150/orig -> origin/gh/guilhermeleobas/150/orig 2025-09-07T07:28:18.7541585Z * [new branch] gh/guilhermeleobas/163/base -> origin/gh/guilhermeleobas/163/base 2025-09-07T07:28:18.7542417Z * [new branch] gh/guilhermeleobas/163/head -> origin/gh/guilhermeleobas/163/head 2025-09-07T07:28:18.7543279Z * [new branch] gh/guilhermeleobas/163/orig -> origin/gh/guilhermeleobas/163/orig 2025-09-07T07:28:18.7544429Z * [new branch] gh/guilhermeleobas/164/base -> origin/gh/guilhermeleobas/164/base 2025-09-07T07:28:18.7545321Z * [new branch] gh/guilhermeleobas/164/head -> origin/gh/guilhermeleobas/164/head 2025-09-07T07:28:18.7546134Z * [new branch] gh/guilhermeleobas/164/orig -> origin/gh/guilhermeleobas/164/orig 2025-09-07T07:28:18.7547367Z * [new branch] gh/guilhermeleobas/165/base -> origin/gh/guilhermeleobas/165/base 2025-09-07T07:28:18.7548342Z * [new branch] gh/guilhermeleobas/165/head -> origin/gh/guilhermeleobas/165/head 2025-09-07T07:28:18.7549201Z * [new branch] gh/guilhermeleobas/165/orig -> origin/gh/guilhermeleobas/165/orig 2025-09-07T07:28:18.7550349Z * [new branch] gh/guilhermeleobas/166/base -> origin/gh/guilhermeleobas/166/base 2025-09-07T07:28:18.7551162Z * [new branch] gh/guilhermeleobas/166/head -> origin/gh/guilhermeleobas/166/head 2025-09-07T07:28:18.7552025Z * [new branch] gh/guilhermeleobas/166/orig -> origin/gh/guilhermeleobas/166/orig 2025-09-07T07:28:18.7553325Z * [new branch] gh/guilhermeleobas/167/base -> origin/gh/guilhermeleobas/167/base 2025-09-07T07:28:18.7554074Z * [new branch] gh/guilhermeleobas/167/head -> origin/gh/guilhermeleobas/167/head 2025-09-07T07:28:18.7554883Z * [new branch] gh/guilhermeleobas/167/orig -> origin/gh/guilhermeleobas/167/orig 2025-09-07T07:28:18.7556024Z * [new branch] gh/guilhermeleobas/168/base -> origin/gh/guilhermeleobas/168/base 2025-09-07T07:28:18.7557116Z * [new branch] gh/guilhermeleobas/168/head -> origin/gh/guilhermeleobas/168/head 2025-09-07T07:28:18.7557708Z * [new branch] gh/guilhermeleobas/168/orig -> origin/gh/guilhermeleobas/168/orig 2025-09-07T07:28:18.7558950Z * [new branch] gh/guilhermeleobas/169/base -> origin/gh/guilhermeleobas/169/base 2025-09-07T07:28:18.7559862Z * [new branch] gh/guilhermeleobas/169/head -> origin/gh/guilhermeleobas/169/head 2025-09-07T07:28:18.7560697Z * [new branch] gh/guilhermeleobas/169/orig -> origin/gh/guilhermeleobas/169/orig 2025-09-07T07:28:18.7561790Z * [new branch] gh/guilhermeleobas/170/base -> origin/gh/guilhermeleobas/170/base 2025-09-07T07:28:18.7562612Z * [new branch] gh/guilhermeleobas/170/head -> origin/gh/guilhermeleobas/170/head 2025-09-07T07:28:18.7563492Z * [new branch] gh/guilhermeleobas/170/orig -> origin/gh/guilhermeleobas/170/orig 2025-09-07T07:28:18.7564600Z * [new branch] gh/guilhermeleobas/171/base -> origin/gh/guilhermeleobas/171/base 2025-09-07T07:28:18.7565573Z * [new branch] gh/guilhermeleobas/171/head -> origin/gh/guilhermeleobas/171/head 2025-09-07T07:28:18.7569088Z * [new branch] gh/guilhermeleobas/171/orig -> origin/gh/guilhermeleobas/171/orig 2025-09-07T07:28:18.7570245Z * [new branch] gh/guilhermeleobas/173/base -> origin/gh/guilhermeleobas/173/base 2025-09-07T07:28:18.7571083Z * [new branch] gh/guilhermeleobas/173/head -> origin/gh/guilhermeleobas/173/head 2025-09-07T07:28:18.7572041Z * [new branch] gh/guilhermeleobas/173/orig -> origin/gh/guilhermeleobas/173/orig 2025-09-07T07:28:18.7573265Z * [new branch] gh/guilhermeleobas/192/base -> origin/gh/guilhermeleobas/192/base 2025-09-07T07:28:18.7574189Z * [new branch] gh/guilhermeleobas/192/head -> origin/gh/guilhermeleobas/192/head 2025-09-07T07:28:18.7575056Z * [new branch] gh/guilhermeleobas/192/orig -> origin/gh/guilhermeleobas/192/orig 2025-09-07T07:28:18.7576583Z * [new branch] gh/guilhermeleobas/193/base -> origin/gh/guilhermeleobas/193/base 2025-09-07T07:28:18.7577440Z * [new branch] gh/guilhermeleobas/193/head -> origin/gh/guilhermeleobas/193/head 2025-09-07T07:28:18.7578407Z * [new branch] gh/guilhermeleobas/193/orig -> origin/gh/guilhermeleobas/193/orig 2025-09-07T07:28:18.7579600Z * [new branch] gh/guilhermeleobas/194/base -> origin/gh/guilhermeleobas/194/base 2025-09-07T07:28:18.7580609Z * [new branch] gh/guilhermeleobas/194/head -> origin/gh/guilhermeleobas/194/head 2025-09-07T07:28:18.7581418Z * [new branch] gh/guilhermeleobas/194/orig -> origin/gh/guilhermeleobas/194/orig 2025-09-07T07:28:18.7582674Z * [new branch] gh/guilhermeleobas/203/base -> origin/gh/guilhermeleobas/203/base 2025-09-07T07:28:18.7583515Z * [new branch] gh/guilhermeleobas/203/head -> origin/gh/guilhermeleobas/203/head 2025-09-07T07:28:18.7584317Z * [new branch] gh/guilhermeleobas/203/orig -> origin/gh/guilhermeleobas/203/orig 2025-09-07T07:28:18.7585614Z * [new branch] gh/guilhermeleobas/204/base -> origin/gh/guilhermeleobas/204/base 2025-09-07T07:28:18.7586447Z * [new branch] gh/guilhermeleobas/204/head -> origin/gh/guilhermeleobas/204/head 2025-09-07T07:28:18.7587328Z * [new branch] gh/guilhermeleobas/204/orig -> origin/gh/guilhermeleobas/204/orig 2025-09-07T07:28:18.7588648Z * [new branch] gh/guilhermeleobas/205/base -> origin/gh/guilhermeleobas/205/base 2025-09-07T07:28:18.7589430Z * [new branch] gh/guilhermeleobas/205/head -> origin/gh/guilhermeleobas/205/head 2025-09-07T07:28:18.7590273Z * [new branch] gh/guilhermeleobas/205/orig -> origin/gh/guilhermeleobas/205/orig 2025-09-07T07:28:18.7591519Z * [new branch] gh/guilhermeleobas/209/base -> origin/gh/guilhermeleobas/209/base 2025-09-07T07:28:18.7592408Z * [new branch] gh/guilhermeleobas/209/head -> origin/gh/guilhermeleobas/209/head 2025-09-07T07:28:18.7593310Z * [new branch] gh/guilhermeleobas/209/orig -> origin/gh/guilhermeleobas/209/orig 2025-09-07T07:28:18.7594509Z * [new branch] gh/guilhermeleobas/210/base -> origin/gh/guilhermeleobas/210/base 2025-09-07T07:28:18.7595357Z * [new branch] gh/guilhermeleobas/210/head -> origin/gh/guilhermeleobas/210/head 2025-09-07T07:28:18.7596234Z * [new branch] gh/guilhermeleobas/210/orig -> origin/gh/guilhermeleobas/210/orig 2025-09-07T07:28:18.7597553Z * [new branch] gh/guilhermeleobas/211/base -> origin/gh/guilhermeleobas/211/base 2025-09-07T07:28:18.7598375Z * [new branch] gh/guilhermeleobas/211/head -> origin/gh/guilhermeleobas/211/head 2025-09-07T07:28:18.7599224Z * [new branch] gh/guilhermeleobas/211/orig -> origin/gh/guilhermeleobas/211/orig 2025-09-07T07:28:18.7600415Z * [new branch] gh/guilhermeleobas/214/base -> origin/gh/guilhermeleobas/214/base 2025-09-07T07:28:18.7601298Z * [new branch] gh/guilhermeleobas/214/head -> origin/gh/guilhermeleobas/214/head 2025-09-07T07:28:18.7602085Z * [new branch] gh/guilhermeleobas/214/orig -> origin/gh/guilhermeleobas/214/orig 2025-09-07T07:28:18.7603324Z * [new branch] gh/guilhermeleobas/215/base -> origin/gh/guilhermeleobas/215/base 2025-09-07T07:28:18.7604180Z * [new branch] gh/guilhermeleobas/215/head -> origin/gh/guilhermeleobas/215/head 2025-09-07T07:28:18.7605017Z * [new branch] gh/guilhermeleobas/215/orig -> origin/gh/guilhermeleobas/215/orig 2025-09-07T07:28:18.7606172Z * [new branch] gh/guilhermeleobas/216/base -> origin/gh/guilhermeleobas/216/base 2025-09-07T07:28:18.7607032Z * [new branch] gh/guilhermeleobas/216/head -> origin/gh/guilhermeleobas/216/head 2025-09-07T07:28:18.7607847Z * [new branch] gh/guilhermeleobas/216/orig -> origin/gh/guilhermeleobas/216/orig 2025-09-07T07:28:18.7609259Z * [new branch] gh/guilhermeleobas/217/base -> origin/gh/guilhermeleobas/217/base 2025-09-07T07:28:18.7610130Z * [new branch] gh/guilhermeleobas/217/head -> origin/gh/guilhermeleobas/217/head 2025-09-07T07:28:18.7610960Z * [new branch] gh/guilhermeleobas/217/orig -> origin/gh/guilhermeleobas/217/orig 2025-09-07T07:28:18.7612195Z * [new branch] gh/guilhermeleobas/219/base -> origin/gh/guilhermeleobas/219/base 2025-09-07T07:28:18.7613028Z * [new branch] gh/guilhermeleobas/219/head -> origin/gh/guilhermeleobas/219/head 2025-09-07T07:28:18.7613842Z * [new branch] gh/guilhermeleobas/219/orig -> origin/gh/guilhermeleobas/219/orig 2025-09-07T07:28:18.7615045Z * [new branch] gh/guilhermeleobas/220/base -> origin/gh/guilhermeleobas/220/base 2025-09-07T07:28:18.7615901Z * [new branch] gh/guilhermeleobas/220/head -> origin/gh/guilhermeleobas/220/head 2025-09-07T07:28:18.7616759Z * [new branch] gh/guilhermeleobas/220/orig -> origin/gh/guilhermeleobas/220/orig 2025-09-07T07:28:18.7617972Z * [new branch] gh/guilhermeleobas/221/base -> origin/gh/guilhermeleobas/221/base 2025-09-07T07:28:18.7618803Z * [new branch] gh/guilhermeleobas/221/head -> origin/gh/guilhermeleobas/221/head 2025-09-07T07:28:18.7619646Z * [new branch] gh/guilhermeleobas/221/orig -> origin/gh/guilhermeleobas/221/orig 2025-09-07T07:28:18.7621115Z * [new branch] gh/guilhermeleobas/222/base -> origin/gh/guilhermeleobas/222/base 2025-09-07T07:28:18.7621813Z * [new branch] gh/guilhermeleobas/222/head -> origin/gh/guilhermeleobas/222/head 2025-09-07T07:28:18.7622681Z * [new branch] gh/guilhermeleobas/222/orig -> origin/gh/guilhermeleobas/222/orig 2025-09-07T07:28:18.7623877Z * [new branch] gh/guilhermeleobas/223/base -> origin/gh/guilhermeleobas/223/base 2025-09-07T07:28:18.7624774Z * [new branch] gh/guilhermeleobas/223/head -> origin/gh/guilhermeleobas/223/head 2025-09-07T07:28:18.7625675Z * [new branch] gh/guilhermeleobas/223/orig -> origin/gh/guilhermeleobas/223/orig 2025-09-07T07:28:18.7626882Z * [new branch] gh/guilhermeleobas/224/base -> origin/gh/guilhermeleobas/224/base 2025-09-07T07:28:18.7627724Z * [new branch] gh/guilhermeleobas/224/head -> origin/gh/guilhermeleobas/224/head 2025-09-07T07:28:18.7628539Z * [new branch] gh/guilhermeleobas/224/orig -> origin/gh/guilhermeleobas/224/orig 2025-09-07T07:28:18.7629706Z * [new branch] gh/guilhermeleobas/225/base -> origin/gh/guilhermeleobas/225/base 2025-09-07T07:28:18.7630601Z * [new branch] gh/guilhermeleobas/225/head -> origin/gh/guilhermeleobas/225/head 2025-09-07T07:28:18.7631452Z * [new branch] gh/guilhermeleobas/225/orig -> origin/gh/guilhermeleobas/225/orig 2025-09-07T07:28:18.7632643Z * [new branch] gh/guilhermeleobas/226/base -> origin/gh/guilhermeleobas/226/base 2025-09-07T07:28:18.7633468Z * [new branch] gh/guilhermeleobas/226/head -> origin/gh/guilhermeleobas/226/head 2025-09-07T07:28:18.7634293Z * [new branch] gh/guilhermeleobas/226/orig -> origin/gh/guilhermeleobas/226/orig 2025-09-07T07:28:18.7635578Z * [new branch] gh/guilhermeleobas/227/base -> origin/gh/guilhermeleobas/227/base 2025-09-07T07:28:18.7636405Z * [new branch] gh/guilhermeleobas/227/head -> origin/gh/guilhermeleobas/227/head 2025-09-07T07:28:18.7637287Z * [new branch] gh/guilhermeleobas/227/orig -> origin/gh/guilhermeleobas/227/orig 2025-09-07T07:28:18.7638528Z * [new branch] gh/guilhermeleobas/228/base -> origin/gh/guilhermeleobas/228/base 2025-09-07T07:28:18.7639428Z * [new branch] gh/guilhermeleobas/228/head -> origin/gh/guilhermeleobas/228/head 2025-09-07T07:28:18.7640169Z * [new branch] gh/guilhermeleobas/228/orig -> origin/gh/guilhermeleobas/228/orig 2025-09-07T07:28:18.7641370Z * [new branch] gh/guilhermeleobas/229/base -> origin/gh/guilhermeleobas/229/base 2025-09-07T07:28:18.7642230Z * [new branch] gh/guilhermeleobas/229/head -> origin/gh/guilhermeleobas/229/head 2025-09-07T07:28:18.7643127Z * [new branch] gh/guilhermeleobas/229/orig -> origin/gh/guilhermeleobas/229/orig 2025-09-07T07:28:18.7644387Z * [new branch] gh/guilhermeleobas/230/base -> origin/gh/guilhermeleobas/230/base 2025-09-07T07:28:18.7645300Z * [new branch] gh/guilhermeleobas/230/head -> origin/gh/guilhermeleobas/230/head 2025-09-07T07:28:18.7646137Z * [new branch] gh/guilhermeleobas/230/orig -> origin/gh/guilhermeleobas/230/orig 2025-09-07T07:28:18.7647334Z * [new branch] gh/guilhermeleobas/231/base -> origin/gh/guilhermeleobas/231/base 2025-09-07T07:28:18.7648147Z * [new branch] gh/guilhermeleobas/231/head -> origin/gh/guilhermeleobas/231/head 2025-09-07T07:28:18.7649016Z * [new branch] gh/guilhermeleobas/231/orig -> origin/gh/guilhermeleobas/231/orig 2025-09-07T07:28:18.7650210Z * [new branch] gh/guilhermeleobas/232/base -> origin/gh/guilhermeleobas/232/base 2025-09-07T07:28:18.7651094Z * [new branch] gh/guilhermeleobas/232/head -> origin/gh/guilhermeleobas/232/head 2025-09-07T07:28:18.7651928Z * [new branch] gh/guilhermeleobas/232/orig -> origin/gh/guilhermeleobas/232/orig 2025-09-07T07:28:18.7653244Z * [new branch] gh/guilhermeleobas/233/base -> origin/gh/guilhermeleobas/233/base 2025-09-07T07:28:18.7653907Z * [new branch] gh/guilhermeleobas/233/head -> origin/gh/guilhermeleobas/233/head 2025-09-07T07:28:18.7654818Z * [new branch] gh/guilhermeleobas/233/orig -> origin/gh/guilhermeleobas/233/orig 2025-09-07T07:28:18.7656065Z * [new branch] gh/guilhermeleobas/234/base -> origin/gh/guilhermeleobas/234/base 2025-09-07T07:28:18.7656904Z * [new branch] gh/guilhermeleobas/234/head -> origin/gh/guilhermeleobas/234/head 2025-09-07T07:28:18.7657764Z * [new branch] gh/guilhermeleobas/234/orig -> origin/gh/guilhermeleobas/234/orig 2025-09-07T07:28:18.7658936Z * [new branch] gh/guilhermeleobas/235/base -> origin/gh/guilhermeleobas/235/base 2025-09-07T07:28:18.7659784Z * [new branch] gh/guilhermeleobas/235/head -> origin/gh/guilhermeleobas/235/head 2025-09-07T07:28:18.7660722Z * [new branch] gh/guilhermeleobas/235/orig -> origin/gh/guilhermeleobas/235/orig 2025-09-07T07:28:18.7661919Z * [new branch] gh/guilhermeleobas/236/base -> origin/gh/guilhermeleobas/236/base 2025-09-07T07:28:18.7662799Z * [new branch] gh/guilhermeleobas/236/head -> origin/gh/guilhermeleobas/236/head 2025-09-07T07:28:18.7663604Z * [new branch] gh/guilhermeleobas/236/orig -> origin/gh/guilhermeleobas/236/orig 2025-09-07T07:28:18.7664751Z * [new branch] gh/guilhermeleobas/237/base -> origin/gh/guilhermeleobas/237/base 2025-09-07T07:28:18.7665620Z * [new branch] gh/guilhermeleobas/237/head -> origin/gh/guilhermeleobas/237/head 2025-09-07T07:28:18.7666581Z * [new branch] gh/guilhermeleobas/237/orig -> origin/gh/guilhermeleobas/237/orig 2025-09-07T07:28:18.7667853Z * [new branch] gh/guilhermeleobas/238/base -> origin/gh/guilhermeleobas/238/base 2025-09-07T07:28:18.7668746Z * [new branch] gh/guilhermeleobas/238/head -> origin/gh/guilhermeleobas/238/head 2025-09-07T07:28:18.7669606Z * [new branch] gh/guilhermeleobas/238/orig -> origin/gh/guilhermeleobas/238/orig 2025-09-07T07:28:18.7670857Z * [new branch] gh/guilhermeleobas/239/base -> origin/gh/guilhermeleobas/239/base 2025-09-07T07:28:18.7671690Z * [new branch] gh/guilhermeleobas/239/head -> origin/gh/guilhermeleobas/239/head 2025-09-07T07:28:18.7672573Z * [new branch] gh/guilhermeleobas/239/orig -> origin/gh/guilhermeleobas/239/orig 2025-09-07T07:28:18.7673858Z * [new branch] gh/guilhermeleobas/240/base -> origin/gh/guilhermeleobas/240/base 2025-09-07T07:28:18.7674649Z * [new branch] gh/guilhermeleobas/240/head -> origin/gh/guilhermeleobas/240/head 2025-09-07T07:28:18.7675492Z * [new branch] gh/guilhermeleobas/240/orig -> origin/gh/guilhermeleobas/240/orig 2025-09-07T07:28:18.7676717Z * [new branch] gh/guilhermeleobas/241/base -> origin/gh/guilhermeleobas/241/base 2025-09-07T07:28:18.7677604Z * [new branch] gh/guilhermeleobas/241/head -> origin/gh/guilhermeleobas/241/head 2025-09-07T07:28:18.7678424Z * [new branch] gh/guilhermeleobas/241/orig -> origin/gh/guilhermeleobas/241/orig 2025-09-07T07:28:18.7679837Z * [new branch] gh/guilhermeleobas/242/base -> origin/gh/guilhermeleobas/242/base 2025-09-07T07:28:18.7680729Z * [new branch] gh/guilhermeleobas/242/head -> origin/gh/guilhermeleobas/242/head 2025-09-07T07:28:18.7681558Z * [new branch] gh/guilhermeleobas/242/orig -> origin/gh/guilhermeleobas/242/orig 2025-09-07T07:28:18.7682706Z * [new branch] gh/guilhermeleobas/243/base -> origin/gh/guilhermeleobas/243/base 2025-09-07T07:28:18.7683536Z * [new branch] gh/guilhermeleobas/243/head -> origin/gh/guilhermeleobas/243/head 2025-09-07T07:28:18.7684387Z * [new branch] gh/guilhermeleobas/243/orig -> origin/gh/guilhermeleobas/243/orig 2025-09-07T07:28:18.7685757Z * [new branch] gh/guilhermeleobas/244/base -> origin/gh/guilhermeleobas/244/base 2025-09-07T07:28:18.7686456Z * [new branch] gh/guilhermeleobas/244/head -> origin/gh/guilhermeleobas/244/head 2025-09-07T07:28:18.7687291Z * [new branch] gh/guilhermeleobas/244/orig -> origin/gh/guilhermeleobas/244/orig 2025-09-07T07:28:18.7688505Z * [new branch] gh/guilhermeleobas/245/base -> origin/gh/guilhermeleobas/245/base 2025-09-07T07:28:18.7689319Z * [new branch] gh/guilhermeleobas/245/head -> origin/gh/guilhermeleobas/245/head 2025-09-07T07:28:18.7690203Z * [new branch] gh/guilhermeleobas/245/orig -> origin/gh/guilhermeleobas/245/orig 2025-09-07T07:28:18.7691459Z * [new branch] gh/guilhermeleobas/73/base -> origin/gh/guilhermeleobas/73/base 2025-09-07T07:28:18.7692333Z * [new branch] gh/guilhermeleobas/73/head -> origin/gh/guilhermeleobas/73/head 2025-09-07T07:28:18.7693126Z * [new branch] gh/guilhermeleobas/73/orig -> origin/gh/guilhermeleobas/73/orig 2025-09-07T07:28:18.7694576Z * [new branch] gh/henrylhtsang/140/base -> origin/gh/henrylhtsang/140/base 2025-09-07T07:28:18.7695670Z * [new branch] gh/henrylhtsang/140/head -> origin/gh/henrylhtsang/140/head 2025-09-07T07:28:18.7696524Z * [new branch] gh/henrylhtsang/140/orig -> origin/gh/henrylhtsang/140/orig 2025-09-07T07:28:18.7697937Z * [new branch] gh/henrylhtsang/141/base -> origin/gh/henrylhtsang/141/base 2025-09-07T07:28:18.7698774Z * [new branch] gh/henrylhtsang/141/head -> origin/gh/henrylhtsang/141/head 2025-09-07T07:28:18.7699652Z * [new branch] gh/henrylhtsang/141/orig -> origin/gh/henrylhtsang/141/orig 2025-09-07T07:28:18.7701035Z * [new branch] gh/henrylhtsang/142/base -> origin/gh/henrylhtsang/142/base 2025-09-07T07:28:18.7702029Z * [new branch] gh/henrylhtsang/142/head -> origin/gh/henrylhtsang/142/head 2025-09-07T07:28:18.7702974Z * [new branch] gh/henrylhtsang/142/orig -> origin/gh/henrylhtsang/142/orig 2025-09-07T07:28:18.7704139Z * [new branch] gh/henrylhtsang/143/base -> origin/gh/henrylhtsang/143/base 2025-09-07T07:28:18.7704983Z * [new branch] gh/henrylhtsang/143/head -> origin/gh/henrylhtsang/143/head 2025-09-07T07:28:18.7705873Z * [new branch] gh/henrylhtsang/143/orig -> origin/gh/henrylhtsang/143/orig 2025-09-07T07:28:18.7707047Z * [new branch] gh/henrylhtsang/144/base -> origin/gh/henrylhtsang/144/base 2025-09-07T07:28:18.7707901Z * [new branch] gh/henrylhtsang/144/head -> origin/gh/henrylhtsang/144/head 2025-09-07T07:28:18.7708814Z * [new branch] gh/henrylhtsang/144/orig -> origin/gh/henrylhtsang/144/orig 2025-09-07T07:28:18.7710250Z * [new branch] gh/henrylhtsang/145/base -> origin/gh/henrylhtsang/145/base 2025-09-07T07:28:18.7711165Z * [new branch] gh/henrylhtsang/145/head -> origin/gh/henrylhtsang/145/head 2025-09-07T07:28:18.7711984Z * [new branch] gh/henrylhtsang/145/orig -> origin/gh/henrylhtsang/145/orig 2025-09-07T07:28:18.7713256Z * [new branch] gh/henrylhtsang/146/base -> origin/gh/henrylhtsang/146/base 2025-09-07T07:28:18.7714326Z * [new branch] gh/henrylhtsang/146/head -> origin/gh/henrylhtsang/146/head 2025-09-07T07:28:18.7715253Z * [new branch] gh/henrylhtsang/146/orig -> origin/gh/henrylhtsang/146/orig 2025-09-07T07:28:18.7716469Z * [new branch] gh/henrylhtsang/147/base -> origin/gh/henrylhtsang/147/base 2025-09-07T07:28:18.7717349Z * [new branch] gh/henrylhtsang/147/head -> origin/gh/henrylhtsang/147/head 2025-09-07T07:28:18.7718209Z * [new branch] gh/henrylhtsang/147/orig -> origin/gh/henrylhtsang/147/orig 2025-09-07T07:28:18.7719616Z * [new branch] gh/henrylhtsang/148/base -> origin/gh/henrylhtsang/148/base 2025-09-07T07:28:18.7720637Z * [new branch] gh/henrylhtsang/148/head -> origin/gh/henrylhtsang/148/head 2025-09-07T07:28:18.7721448Z * [new branch] gh/henrylhtsang/148/orig -> origin/gh/henrylhtsang/148/orig 2025-09-07T07:28:18.7722599Z * [new branch] gh/henrylhtsang/149/base -> origin/gh/henrylhtsang/149/base 2025-09-07T07:28:18.7723536Z * [new branch] gh/henrylhtsang/149/head -> origin/gh/henrylhtsang/149/head 2025-09-07T07:28:18.7724392Z * [new branch] gh/henrylhtsang/149/orig -> origin/gh/henrylhtsang/149/orig 2025-09-07T07:28:18.7725823Z * [new branch] gh/huydhn/1/next -> origin/gh/huydhn/1/next 2025-09-07T07:28:18.7726991Z * [new branch] gh/huydhn/2/next -> origin/gh/huydhn/2/next 2025-09-07T07:28:18.7728157Z * [new branch] gh/huydhn/3/next -> origin/gh/huydhn/3/next 2025-09-07T07:28:18.7729400Z * [new branch] gh/huydhn/4/next -> origin/gh/huydhn/4/next 2025-09-07T07:28:18.7730514Z * [new branch] gh/huydhn/5/next -> origin/gh/huydhn/5/next 2025-09-07T07:28:18.7731597Z * [new branch] gh/huydhn/6/next -> origin/gh/huydhn/6/next 2025-09-07T07:28:18.7733078Z * [new branch] gh/int3/97/base -> origin/gh/int3/97/base 2025-09-07T07:28:18.7733922Z * [new branch] gh/int3/97/head -> origin/gh/int3/97/head 2025-09-07T07:28:18.7735429Z * [new branch] gh/isuruf/101/base -> origin/gh/isuruf/101/base 2025-09-07T07:28:18.7736271Z * [new branch] gh/isuruf/101/head -> origin/gh/isuruf/101/head 2025-09-07T07:28:18.7737460Z * [new branch] gh/isuruf/141/base -> origin/gh/isuruf/141/base 2025-09-07T07:28:18.7738303Z * [new branch] gh/isuruf/141/head -> origin/gh/isuruf/141/head 2025-09-07T07:28:18.7739151Z * [new branch] gh/isuruf/141/orig -> origin/gh/isuruf/141/orig 2025-09-07T07:28:18.7740332Z * [new branch] gh/isuruf/142/base -> origin/gh/isuruf/142/base 2025-09-07T07:28:18.7741154Z * [new branch] gh/isuruf/142/head -> origin/gh/isuruf/142/head 2025-09-07T07:28:18.7742043Z * [new branch] gh/isuruf/142/orig -> origin/gh/isuruf/142/orig 2025-09-07T07:28:18.7743173Z * [new branch] gh/isuruf/143/base -> origin/gh/isuruf/143/base 2025-09-07T07:28:18.7744045Z * [new branch] gh/isuruf/143/head -> origin/gh/isuruf/143/head 2025-09-07T07:28:18.7744835Z * [new branch] gh/isuruf/143/orig -> origin/gh/isuruf/143/orig 2025-09-07T07:28:18.7746021Z * [new branch] gh/isuruf/144/base -> origin/gh/isuruf/144/base 2025-09-07T07:28:18.7747169Z * [new branch] gh/isuruf/144/head -> origin/gh/isuruf/144/head 2025-09-07T07:28:18.7748031Z * [new branch] gh/isuruf/144/orig -> origin/gh/isuruf/144/orig 2025-09-07T07:28:18.7749188Z * [new branch] gh/isuruf/145/base -> origin/gh/isuruf/145/base 2025-09-07T07:28:18.7749973Z * [new branch] gh/isuruf/145/head -> origin/gh/isuruf/145/head 2025-09-07T07:28:18.7750892Z * [new branch] gh/isuruf/145/orig -> origin/gh/isuruf/145/orig 2025-09-07T07:28:18.7752067Z * [new branch] gh/isuruf/146/base -> origin/gh/isuruf/146/base 2025-09-07T07:28:18.7752896Z * [new branch] gh/isuruf/146/head -> origin/gh/isuruf/146/head 2025-09-07T07:28:18.7753809Z * [new branch] gh/isuruf/146/orig -> origin/gh/isuruf/146/orig 2025-09-07T07:28:18.7755051Z * [new branch] gh/isuruf/81/base -> origin/gh/isuruf/81/base 2025-09-07T07:28:18.7755919Z * [new branch] gh/isuruf/81/head -> origin/gh/isuruf/81/head 2025-09-07T07:28:18.7756873Z * [new branch] gh/isuruf/81/orig -> origin/gh/isuruf/81/orig 2025-09-07T07:28:18.7758321Z * [new branch] gh/jamesjwu/150/base -> origin/gh/jamesjwu/150/base 2025-09-07T07:28:18.7759150Z * [new branch] gh/jamesjwu/150/head -> origin/gh/jamesjwu/150/head 2025-09-07T07:28:18.7759998Z * [new branch] gh/jamesjwu/150/orig -> origin/gh/jamesjwu/150/orig 2025-09-07T07:28:18.7761286Z * [new branch] gh/jamesjwu/154/base -> origin/gh/jamesjwu/154/base 2025-09-07T07:28:18.7762096Z * [new branch] gh/jamesjwu/154/head -> origin/gh/jamesjwu/154/head 2025-09-07T07:28:18.7762908Z * [new branch] gh/jamesjwu/154/orig -> origin/gh/jamesjwu/154/orig 2025-09-07T07:28:18.7764098Z * [new branch] gh/jamesjwu/155/base -> origin/gh/jamesjwu/155/base 2025-09-07T07:28:18.7764967Z * [new branch] gh/jamesjwu/155/head -> origin/gh/jamesjwu/155/head 2025-09-07T07:28:18.7765943Z * [new branch] gh/jamesjwu/155/orig -> origin/gh/jamesjwu/155/orig 2025-09-07T07:28:18.7767063Z * [new branch] gh/jamesjwu/159/base -> origin/gh/jamesjwu/159/base 2025-09-07T07:28:18.7767939Z * [new branch] gh/jamesjwu/159/head -> origin/gh/jamesjwu/159/head 2025-09-07T07:28:18.7768818Z * [new branch] gh/jamesjwu/159/orig -> origin/gh/jamesjwu/159/orig 2025-09-07T07:28:18.7770322Z * [new branch] gh/jamesjwu/163/base -> origin/gh/jamesjwu/163/base 2025-09-07T07:28:18.7771192Z * [new branch] gh/jamesjwu/163/head -> origin/gh/jamesjwu/163/head 2025-09-07T07:28:18.7772047Z * [new branch] gh/jamesjwu/163/orig -> origin/gh/jamesjwu/163/orig 2025-09-07T07:28:18.7773186Z * [new branch] gh/jamesjwu/171/base -> origin/gh/jamesjwu/171/base 2025-09-07T07:28:18.7774028Z * [new branch] gh/jamesjwu/171/head -> origin/gh/jamesjwu/171/head 2025-09-07T07:28:18.7774853Z * [new branch] gh/jamesjwu/171/orig -> origin/gh/jamesjwu/171/orig 2025-09-07T07:28:18.7776013Z * [new branch] gh/jamesjwu/176/base -> origin/gh/jamesjwu/176/base 2025-09-07T07:28:18.7776845Z * [new branch] gh/jamesjwu/176/head -> origin/gh/jamesjwu/176/head 2025-09-07T07:28:18.7777667Z * [new branch] gh/jamesjwu/176/orig -> origin/gh/jamesjwu/176/orig 2025-09-07T07:28:18.7778848Z * [new branch] gh/jamesjwu/181/base -> origin/gh/jamesjwu/181/base 2025-09-07T07:28:18.7779659Z * [new branch] gh/jamesjwu/181/head -> origin/gh/jamesjwu/181/head 2025-09-07T07:28:18.7780552Z * [new branch] gh/jamesjwu/181/orig -> origin/gh/jamesjwu/181/orig 2025-09-07T07:28:18.7781775Z * [new branch] gh/jamesjwu/182/base -> origin/gh/jamesjwu/182/base 2025-09-07T07:28:18.7782701Z * [new branch] gh/jamesjwu/182/head -> origin/gh/jamesjwu/182/head 2025-09-07T07:28:18.7783525Z * [new branch] gh/jamesjwu/182/orig -> origin/gh/jamesjwu/182/orig 2025-09-07T07:28:18.7784711Z * [new branch] gh/jamesjwu/183/base -> origin/gh/jamesjwu/183/base 2025-09-07T07:28:18.7785541Z * [new branch] gh/jamesjwu/183/head -> origin/gh/jamesjwu/183/head 2025-09-07T07:28:18.7786391Z * [new branch] gh/jamesjwu/183/orig -> origin/gh/jamesjwu/183/orig 2025-09-07T07:28:18.7787536Z * [new branch] gh/jamesjwu/184/base -> origin/gh/jamesjwu/184/base 2025-09-07T07:28:18.7788406Z * [new branch] gh/jamesjwu/184/head -> origin/gh/jamesjwu/184/head 2025-09-07T07:28:18.7789315Z * [new branch] gh/jamesjwu/184/orig -> origin/gh/jamesjwu/184/orig 2025-09-07T07:28:18.7790481Z * [new branch] gh/jamesjwu/185/base -> origin/gh/jamesjwu/185/base 2025-09-07T07:28:18.7791389Z * [new branch] gh/jamesjwu/185/head -> origin/gh/jamesjwu/185/head 2025-09-07T07:28:18.7792122Z * [new branch] gh/jamesjwu/185/orig -> origin/gh/jamesjwu/185/orig 2025-09-07T07:28:18.7793345Z * [new branch] gh/jamesjwu/186/base -> origin/gh/jamesjwu/186/base 2025-09-07T07:28:18.7794166Z * [new branch] gh/jamesjwu/186/head -> origin/gh/jamesjwu/186/head 2025-09-07T07:28:18.7795013Z * [new branch] gh/jamesjwu/186/orig -> origin/gh/jamesjwu/186/orig 2025-09-07T07:28:18.7796199Z * [new branch] gh/jamesjwu/187/base -> origin/gh/jamesjwu/187/base 2025-09-07T07:28:18.7797063Z * [new branch] gh/jamesjwu/187/head -> origin/gh/jamesjwu/187/head 2025-09-07T07:28:18.7798091Z * [new branch] gh/jamesjwu/187/orig -> origin/gh/jamesjwu/187/orig 2025-09-07T07:28:18.7799074Z * [new branch] gh/jamesjwu/188/base -> origin/gh/jamesjwu/188/base 2025-09-07T07:28:18.7799905Z * [new branch] gh/jamesjwu/188/head -> origin/gh/jamesjwu/188/head 2025-09-07T07:28:18.7800695Z * [new branch] gh/jamesjwu/188/orig -> origin/gh/jamesjwu/188/orig 2025-09-07T07:28:18.7801796Z * [new branch] gh/jamesjwu/189/base -> origin/gh/jamesjwu/189/base 2025-09-07T07:28:18.7802701Z * [new branch] gh/jamesjwu/189/head -> origin/gh/jamesjwu/189/head 2025-09-07T07:28:18.7803516Z * [new branch] gh/jamesjwu/189/orig -> origin/gh/jamesjwu/189/orig 2025-09-07T07:28:18.7805048Z * [new branch] gh/jamesjwu/190/base -> origin/gh/jamesjwu/190/base 2025-09-07T07:28:18.7806020Z * [new branch] gh/jamesjwu/190/head -> origin/gh/jamesjwu/190/head 2025-09-07T07:28:18.7806861Z * [new branch] gh/jamesjwu/190/orig -> origin/gh/jamesjwu/190/orig 2025-09-07T07:28:18.7808225Z * [new branch] gh/jamesjwu/52/base -> origin/gh/jamesjwu/52/base 2025-09-07T07:28:18.7809153Z * [new branch] gh/jamesjwu/52/head -> origin/gh/jamesjwu/52/head 2025-09-07T07:28:18.7810316Z * [new branch] gh/jamesjwu/53/base -> origin/gh/jamesjwu/53/base 2025-09-07T07:28:18.7811109Z * [new branch] gh/jamesjwu/53/head -> origin/gh/jamesjwu/53/head 2025-09-07T07:28:18.7812284Z * [new branch] gh/jamesjwu/54/base -> origin/gh/jamesjwu/54/base 2025-09-07T07:28:18.7813113Z * [new branch] gh/jamesjwu/54/head -> origin/gh/jamesjwu/54/head 2025-09-07T07:28:18.7814198Z * [new branch] gh/jamesjwu/55/base -> origin/gh/jamesjwu/55/base 2025-09-07T07:28:18.7815251Z * [new branch] gh/jamesjwu/55/head -> origin/gh/jamesjwu/55/head 2025-09-07T07:28:18.7816416Z * [new branch] gh/jamesjwu/56/base -> origin/gh/jamesjwu/56/base 2025-09-07T07:28:18.7817252Z * [new branch] gh/jamesjwu/56/head -> origin/gh/jamesjwu/56/head 2025-09-07T07:28:18.7818282Z * [new branch] gh/jamesjwu/57/base -> origin/gh/jamesjwu/57/base 2025-09-07T07:28:18.7819154Z * [new branch] gh/jamesjwu/57/head -> origin/gh/jamesjwu/57/head 2025-09-07T07:28:18.7820178Z * [new branch] gh/jamesjwu/58/base -> origin/gh/jamesjwu/58/base 2025-09-07T07:28:18.7821037Z * [new branch] gh/jamesjwu/58/head -> origin/gh/jamesjwu/58/head 2025-09-07T07:28:18.7822134Z * [new branch] gh/jamesjwu/59/base -> origin/gh/jamesjwu/59/base 2025-09-07T07:28:18.7823132Z * [new branch] gh/jamesjwu/59/head -> origin/gh/jamesjwu/59/head 2025-09-07T07:28:18.7824284Z * [new branch] gh/jamesjwu/60/base -> origin/gh/jamesjwu/60/base 2025-09-07T07:28:18.7825108Z * [new branch] gh/jamesjwu/60/head -> origin/gh/jamesjwu/60/head 2025-09-07T07:28:18.7826269Z * [new branch] gh/jamesjwu/61/base -> origin/gh/jamesjwu/61/base 2025-09-07T07:28:18.7827060Z * [new branch] gh/jamesjwu/61/head -> origin/gh/jamesjwu/61/head 2025-09-07T07:28:18.7828159Z * [new branch] gh/jamesjwu/62/base -> origin/gh/jamesjwu/62/base 2025-09-07T07:28:18.7828948Z * [new branch] gh/jamesjwu/62/head -> origin/gh/jamesjwu/62/head 2025-09-07T07:28:18.7830230Z * [new branch] gh/jamesjwu/63/base -> origin/gh/jamesjwu/63/base 2025-09-07T07:28:18.7831202Z * [new branch] gh/jamesjwu/63/head -> origin/gh/jamesjwu/63/head 2025-09-07T07:28:18.7832413Z * [new branch] gh/jamesjwu/64/base -> origin/gh/jamesjwu/64/base 2025-09-07T07:28:18.7833280Z * [new branch] gh/jamesjwu/64/head -> origin/gh/jamesjwu/64/head 2025-09-07T07:28:18.7834450Z * [new branch] gh/jamesjwu/65/base -> origin/gh/jamesjwu/65/base 2025-09-07T07:28:18.7835236Z * [new branch] gh/jamesjwu/65/head -> origin/gh/jamesjwu/65/head 2025-09-07T07:28:18.7837069Z * [new branch] gh/janeyx99/165/base -> origin/gh/janeyx99/165/base 2025-09-07T07:28:18.7838086Z * [new branch] gh/janeyx99/165/head -> origin/gh/janeyx99/165/head 2025-09-07T07:28:18.7838914Z * [new branch] gh/janeyx99/165/orig -> origin/gh/janeyx99/165/orig 2025-09-07T07:28:18.7839955Z * [new branch] gh/janeyx99/201/base -> origin/gh/janeyx99/201/base 2025-09-07T07:28:18.7840799Z * [new branch] gh/janeyx99/201/head -> origin/gh/janeyx99/201/head 2025-09-07T07:28:18.7841693Z * [new branch] gh/janeyx99/201/orig -> origin/gh/janeyx99/201/orig 2025-09-07T07:28:18.7843117Z * [new branch] gh/janeyx99/225/base -> origin/gh/janeyx99/225/base 2025-09-07T07:28:18.7843998Z * [new branch] gh/janeyx99/225/head -> origin/gh/janeyx99/225/head 2025-09-07T07:28:18.7844840Z * [new branch] gh/janeyx99/225/orig -> origin/gh/janeyx99/225/orig 2025-09-07T07:28:18.7846048Z * [new branch] gh/janeyx99/296/base -> origin/gh/janeyx99/296/base 2025-09-07T07:28:18.7846884Z * [new branch] gh/janeyx99/296/head -> origin/gh/janeyx99/296/head 2025-09-07T07:28:18.7847706Z * [new branch] gh/janeyx99/296/orig -> origin/gh/janeyx99/296/orig 2025-09-07T07:28:18.7848871Z * [new branch] gh/janeyx99/297/base -> origin/gh/janeyx99/297/base 2025-09-07T07:28:18.7849768Z * [new branch] gh/janeyx99/297/head -> origin/gh/janeyx99/297/head 2025-09-07T07:28:18.7850634Z * [new branch] gh/janeyx99/297/orig -> origin/gh/janeyx99/297/orig 2025-09-07T07:28:18.7851874Z * [new branch] gh/janeyx99/298/base -> origin/gh/janeyx99/298/base 2025-09-07T07:28:18.7852725Z * [new branch] gh/janeyx99/298/head -> origin/gh/janeyx99/298/head 2025-09-07T07:28:18.7853570Z * [new branch] gh/janeyx99/298/orig -> origin/gh/janeyx99/298/orig 2025-09-07T07:28:18.7854711Z * [new branch] gh/janeyx99/299/base -> origin/gh/janeyx99/299/base 2025-09-07T07:28:18.7855533Z * [new branch] gh/janeyx99/299/head -> origin/gh/janeyx99/299/head 2025-09-07T07:28:18.7856395Z * [new branch] gh/janeyx99/299/orig -> origin/gh/janeyx99/299/orig 2025-09-07T07:28:18.7857938Z * [new branch] gh/janeyx99/300/base -> origin/gh/janeyx99/300/base 2025-09-07T07:28:18.7858994Z * [new branch] gh/janeyx99/300/head -> origin/gh/janeyx99/300/head 2025-09-07T07:28:18.7859912Z * [new branch] gh/janeyx99/300/orig -> origin/gh/janeyx99/300/orig 2025-09-07T07:28:18.7861078Z * [new branch] gh/janeyx99/301/base -> origin/gh/janeyx99/301/base 2025-09-07T07:28:18.7862061Z * [new branch] gh/janeyx99/301/head -> origin/gh/janeyx99/301/head 2025-09-07T07:28:18.7862839Z * [new branch] gh/janeyx99/301/orig -> origin/gh/janeyx99/301/orig 2025-09-07T07:28:18.7863944Z * [new branch] gh/janeyx99/302/base -> origin/gh/janeyx99/302/base 2025-09-07T07:28:18.7864791Z * [new branch] gh/janeyx99/302/head -> origin/gh/janeyx99/302/head 2025-09-07T07:28:18.7865979Z * [new branch] gh/janeyx99/303/base -> origin/gh/janeyx99/303/base 2025-09-07T07:28:18.7866788Z * [new branch] gh/janeyx99/303/head -> origin/gh/janeyx99/303/head 2025-09-07T07:28:18.7868050Z * [new branch] gh/janeyx99/88/base -> origin/gh/janeyx99/88/base 2025-09-07T07:28:18.7868919Z * [new branch] gh/janeyx99/88/head -> origin/gh/janeyx99/88/head 2025-09-07T07:28:18.7869791Z * [new branch] gh/janeyx99/88/orig -> origin/gh/janeyx99/88/orig 2025-09-07T07:28:18.7871347Z * [new branch] gh/jansel/360/base -> origin/gh/jansel/360/base 2025-09-07T07:28:18.7872201Z * [new branch] gh/jansel/360/head -> origin/gh/jansel/360/head 2025-09-07T07:28:18.7873427Z * [new branch] gh/jansel/451/base -> origin/gh/jansel/451/base 2025-09-07T07:28:18.7874333Z * [new branch] gh/jansel/451/head -> origin/gh/jansel/451/head 2025-09-07T07:28:18.7875158Z * [new branch] gh/jansel/451/orig -> origin/gh/jansel/451/orig 2025-09-07T07:28:18.7876319Z * [new branch] gh/jansel/462/base -> origin/gh/jansel/462/base 2025-09-07T07:28:18.7877204Z * [new branch] gh/jansel/462/head -> origin/gh/jansel/462/head 2025-09-07T07:28:18.7878038Z * [new branch] gh/jansel/462/orig -> origin/gh/jansel/462/orig 2025-09-07T07:28:18.7879190Z * [new branch] gh/jansel/531/base -> origin/gh/jansel/531/base 2025-09-07T07:28:18.7880102Z * [new branch] gh/jansel/531/head -> origin/gh/jansel/531/head 2025-09-07T07:28:18.7880955Z * [new branch] gh/jansel/531/orig -> origin/gh/jansel/531/orig 2025-09-07T07:28:18.7882464Z * [new branch] gh/jbschlosser/208/head -> origin/gh/jbschlosser/208/head 2025-09-07T07:28:18.7883828Z * [new branch] gh/jbschlosser/247/base -> origin/gh/jbschlosser/247/base 2025-09-07T07:28:18.7884678Z * [new branch] gh/jbschlosser/247/head -> origin/gh/jbschlosser/247/head 2025-09-07T07:28:18.7885482Z * [new branch] gh/jbschlosser/247/orig -> origin/gh/jbschlosser/247/orig 2025-09-07T07:28:18.7886754Z * [new branch] gh/jbschlosser/248/base -> origin/gh/jbschlosser/248/base 2025-09-07T07:28:18.7887620Z * [new branch] gh/jbschlosser/248/head -> origin/gh/jbschlosser/248/head 2025-09-07T07:28:18.7888476Z * [new branch] gh/jbschlosser/248/orig -> origin/gh/jbschlosser/248/orig 2025-09-07T07:28:18.7889727Z * [new branch] gh/jbschlosser/250/base -> origin/gh/jbschlosser/250/base 2025-09-07T07:28:18.7890557Z * [new branch] gh/jbschlosser/250/head -> origin/gh/jbschlosser/250/head 2025-09-07T07:28:18.7891370Z * [new branch] gh/jbschlosser/250/orig -> origin/gh/jbschlosser/250/orig 2025-09-07T07:28:18.7892786Z * [new branch] gh/jiayisunx/59/base -> origin/gh/jiayisunx/59/base 2025-09-07T07:28:18.7893653Z * [new branch] gh/jiayisunx/59/head -> origin/gh/jiayisunx/59/head 2025-09-07T07:28:18.7894502Z * [new branch] gh/jiayisunx/59/orig -> origin/gh/jiayisunx/59/orig 2025-09-07T07:28:18.7895672Z * [new branch] gh/jiayisunx/61/base -> origin/gh/jiayisunx/61/base 2025-09-07T07:28:18.7896508Z * [new branch] gh/jiayisunx/61/head -> origin/gh/jiayisunx/61/head 2025-09-07T07:28:18.7897558Z * [new branch] gh/jiayisunx/61/orig -> origin/gh/jiayisunx/61/orig 2025-09-07T07:28:18.7898549Z * [new branch] gh/jiayisunx/64/base -> origin/gh/jiayisunx/64/base 2025-09-07T07:28:18.7899389Z * [new branch] gh/jiayisunx/64/head -> origin/gh/jiayisunx/64/head 2025-09-07T07:28:18.7900306Z * [new branch] gh/jiayisunx/64/orig -> origin/gh/jiayisunx/64/orig 2025-09-07T07:28:18.7901494Z * [new branch] gh/jiayisunx/65/base -> origin/gh/jiayisunx/65/base 2025-09-07T07:28:18.7902572Z * [new branch] gh/jiayisunx/65/head -> origin/gh/jiayisunx/65/head 2025-09-07T07:28:18.7903471Z * [new branch] gh/jiayisunx/65/orig -> origin/gh/jiayisunx/65/orig 2025-09-07T07:28:18.7904579Z * [new branch] gh/jiayisunx/66/base -> origin/gh/jiayisunx/66/base 2025-09-07T07:28:18.7905450Z * [new branch] gh/jiayisunx/66/head -> origin/gh/jiayisunx/66/head 2025-09-07T07:28:18.7906494Z * [new branch] gh/jiayisunx/66/orig -> origin/gh/jiayisunx/66/orig 2025-09-07T07:28:18.7907743Z * [new branch] gh/jiayisunx/67/base -> origin/gh/jiayisunx/67/base 2025-09-07T07:28:18.7908591Z * [new branch] gh/jiayisunx/67/head -> origin/gh/jiayisunx/67/head 2025-09-07T07:28:18.7909518Z * [new branch] gh/jiayisunx/67/orig -> origin/gh/jiayisunx/67/orig 2025-09-07T07:28:18.7910684Z * [new branch] gh/jiayisunx/68/base -> origin/gh/jiayisunx/68/base 2025-09-07T07:28:18.7911514Z * [new branch] gh/jiayisunx/68/head -> origin/gh/jiayisunx/68/head 2025-09-07T07:28:18.7912326Z * [new branch] gh/jiayisunx/68/orig -> origin/gh/jiayisunx/68/orig 2025-09-07T07:28:18.7913535Z * [new branch] gh/jiayisunx/69/base -> origin/gh/jiayisunx/69/base 2025-09-07T07:28:18.7914355Z * [new branch] gh/jiayisunx/69/head -> origin/gh/jiayisunx/69/head 2025-09-07T07:28:18.7915203Z * [new branch] gh/jiayisunx/69/orig -> origin/gh/jiayisunx/69/orig 2025-09-07T07:28:18.7916450Z * [new branch] gh/jiayisunx/70/base -> origin/gh/jiayisunx/70/base 2025-09-07T07:28:18.7917299Z * [new branch] gh/jiayisunx/70/head -> origin/gh/jiayisunx/70/head 2025-09-07T07:28:18.7918114Z * [new branch] gh/jiayisunx/70/orig -> origin/gh/jiayisunx/70/orig 2025-09-07T07:28:18.7919333Z * [new branch] gh/jiayisunx/71/base -> origin/gh/jiayisunx/71/base 2025-09-07T07:28:18.7920152Z * [new branch] gh/jiayisunx/71/head -> origin/gh/jiayisunx/71/head 2025-09-07T07:28:18.7921045Z * [new branch] gh/jiayisunx/71/orig -> origin/gh/jiayisunx/71/orig 2025-09-07T07:28:18.7922199Z * [new branch] gh/jiayisunx/72/base -> origin/gh/jiayisunx/72/base 2025-09-07T07:28:18.7923034Z * [new branch] gh/jiayisunx/72/head -> origin/gh/jiayisunx/72/head 2025-09-07T07:28:18.7923877Z * [new branch] gh/jiayisunx/72/orig -> origin/gh/jiayisunx/72/orig 2025-09-07T07:28:18.7925038Z * [new branch] gh/jiayisunx/73/base -> origin/gh/jiayisunx/73/base 2025-09-07T07:28:18.7925917Z * [new branch] gh/jiayisunx/73/head -> origin/gh/jiayisunx/73/head 2025-09-07T07:28:18.7926764Z * [new branch] gh/jiayisunx/73/orig -> origin/gh/jiayisunx/73/orig 2025-09-07T07:28:18.7928128Z * [new branch] gh/jiayisunx/74/base -> origin/gh/jiayisunx/74/base 2025-09-07T07:28:18.7928987Z * [new branch] gh/jiayisunx/74/head -> origin/gh/jiayisunx/74/head 2025-09-07T07:28:18.7929945Z * [new branch] gh/jiayisunx/74/orig -> origin/gh/jiayisunx/74/orig 2025-09-07T07:28:18.7931135Z * [new branch] gh/jiayisunx/75/base -> origin/gh/jiayisunx/75/base 2025-09-07T07:28:18.7932005Z * [new branch] gh/jiayisunx/75/head -> origin/gh/jiayisunx/75/head 2025-09-07T07:28:18.7932815Z * [new branch] gh/jiayisunx/75/orig -> origin/gh/jiayisunx/75/orig 2025-09-07T07:28:18.7933821Z * [new branch] gh/jiayisunx/76/base -> origin/gh/jiayisunx/76/base 2025-09-07T07:28:18.7934582Z * [new branch] gh/jiayisunx/76/head -> origin/gh/jiayisunx/76/head 2025-09-07T07:28:18.7935447Z * [new branch] gh/jiayisunx/76/orig -> origin/gh/jiayisunx/76/orig 2025-09-07T07:28:18.7936847Z * [new branch] gh/jjwu@meta.com/1/base -> origin/gh/jjwu@meta.com/1/base 2025-09-07T07:28:18.7937672Z * [new branch] gh/jjwu@meta.com/1/head -> origin/gh/jjwu@meta.com/1/head 2025-09-07T07:28:18.7939101Z * [new branch] gh/justinchuby/111/base -> origin/gh/justinchuby/111/base 2025-09-07T07:28:18.7940079Z * [new branch] gh/justinchuby/111/head -> origin/gh/justinchuby/111/head 2025-09-07T07:28:18.7941093Z * [new branch] gh/justinchuby/111/orig -> origin/gh/justinchuby/111/orig 2025-09-07T07:28:18.7942287Z * [new branch] gh/justinchuby/112/base -> origin/gh/justinchuby/112/base 2025-09-07T07:28:18.7943074Z * [new branch] gh/justinchuby/112/head -> origin/gh/justinchuby/112/head 2025-09-07T07:28:18.7943981Z * [new branch] gh/justinchuby/112/orig -> origin/gh/justinchuby/112/orig 2025-09-07T07:28:18.7945141Z * [new branch] gh/justinchuby/113/base -> origin/gh/justinchuby/113/base 2025-09-07T07:28:18.7945992Z * [new branch] gh/justinchuby/113/head -> origin/gh/justinchuby/113/head 2025-09-07T07:28:18.7946923Z * [new branch] gh/justinchuby/113/orig -> origin/gh/justinchuby/113/orig 2025-09-07T07:28:18.7947994Z * [new branch] gh/justinchuby/114/base -> origin/gh/justinchuby/114/base 2025-09-07T07:28:18.7948869Z * [new branch] gh/justinchuby/114/head -> origin/gh/justinchuby/114/head 2025-09-07T07:28:18.7949693Z * [new branch] gh/justinchuby/114/orig -> origin/gh/justinchuby/114/orig 2025-09-07T07:28:18.7950857Z * [new branch] gh/justinchuby/115/base -> origin/gh/justinchuby/115/base 2025-09-07T07:28:18.7951722Z * [new branch] gh/justinchuby/115/head -> origin/gh/justinchuby/115/head 2025-09-07T07:28:18.7952765Z * [new branch] gh/justinchuby/115/orig -> origin/gh/justinchuby/115/orig 2025-09-07T07:28:18.7954197Z * [new branch] gh/karthickai/1/base -> origin/gh/karthickai/1/base 2025-09-07T07:28:18.7955077Z * [new branch] gh/karthickai/1/head -> origin/gh/karthickai/1/head 2025-09-07T07:28:18.7955999Z * [new branch] gh/karthickai/1/orig -> origin/gh/karthickai/1/orig 2025-09-07T07:28:18.7957186Z * [new branch] gh/karthickai/2/base -> origin/gh/karthickai/2/base 2025-09-07T07:28:18.7958056Z * [new branch] gh/karthickai/2/head -> origin/gh/karthickai/2/head 2025-09-07T07:28:18.7958832Z * [new branch] gh/karthickai/2/orig -> origin/gh/karthickai/2/orig 2025-09-07T07:28:18.7960243Z * [new branch] gh/kurtamohler/32/base -> origin/gh/kurtamohler/32/base 2025-09-07T07:28:18.7961123Z * [new branch] gh/kurtamohler/32/head -> origin/gh/kurtamohler/32/head 2025-09-07T07:28:18.7961958Z * [new branch] gh/kurtamohler/32/orig -> origin/gh/kurtamohler/32/orig 2025-09-07T07:28:18.7963138Z * [new branch] gh/kurtamohler/33/base -> origin/gh/kurtamohler/33/base 2025-09-07T07:28:18.7964026Z * [new branch] gh/kurtamohler/33/head -> origin/gh/kurtamohler/33/head 2025-09-07T07:28:18.7964851Z * [new branch] gh/kurtamohler/33/orig -> origin/gh/kurtamohler/33/orig 2025-09-07T07:28:18.7969556Z * [new branch] gh/kurtamohler/34/base -> origin/gh/kurtamohler/34/base 2025-09-07T07:28:18.7970532Z * [new branch] gh/kurtamohler/34/head -> origin/gh/kurtamohler/34/head 2025-09-07T07:28:18.7971291Z * [new branch] gh/kurtamohler/34/orig -> origin/gh/kurtamohler/34/orig 2025-09-07T07:28:18.7972423Z * [new branch] gh/kurtamohler/41/base -> origin/gh/kurtamohler/41/base 2025-09-07T07:28:18.7973270Z * [new branch] gh/kurtamohler/41/head -> origin/gh/kurtamohler/41/head 2025-09-07T07:28:18.7974115Z * [new branch] gh/kurtamohler/41/orig -> origin/gh/kurtamohler/41/orig 2025-09-07T07:28:18.7975293Z * [new branch] gh/kurtamohler/46/base -> origin/gh/kurtamohler/46/base 2025-09-07T07:28:18.7976098Z * [new branch] gh/kurtamohler/46/head -> origin/gh/kurtamohler/46/head 2025-09-07T07:28:18.7976930Z * [new branch] gh/kurtamohler/46/orig -> origin/gh/kurtamohler/46/orig 2025-09-07T07:28:18.7978447Z * [new branch] gh/kurtamohler/47/base -> origin/gh/kurtamohler/47/base 2025-09-07T07:28:18.7979391Z * [new branch] gh/kurtamohler/47/head -> origin/gh/kurtamohler/47/head 2025-09-07T07:28:18.7980240Z * [new branch] gh/kurtamohler/47/orig -> origin/gh/kurtamohler/47/orig 2025-09-07T07:28:18.7981439Z * [new branch] gh/kurtamohler/48/base -> origin/gh/kurtamohler/48/base 2025-09-07T07:28:18.7982243Z * [new branch] gh/kurtamohler/48/head -> origin/gh/kurtamohler/48/head 2025-09-07T07:28:18.7983121Z * [new branch] gh/kurtamohler/48/orig -> origin/gh/kurtamohler/48/orig 2025-09-07T07:28:18.7984307Z * [new branch] gh/kurtamohler/49/base -> origin/gh/kurtamohler/49/base 2025-09-07T07:28:18.7985129Z * [new branch] gh/kurtamohler/49/head -> origin/gh/kurtamohler/49/head 2025-09-07T07:28:18.7985943Z * [new branch] gh/kurtamohler/49/orig -> origin/gh/kurtamohler/49/orig 2025-09-07T07:28:18.7987146Z * [new branch] gh/kurtamohler/50/base -> origin/gh/kurtamohler/50/base 2025-09-07T07:28:18.7987941Z * [new branch] gh/kurtamohler/50/head -> origin/gh/kurtamohler/50/head 2025-09-07T07:28:18.7988814Z * [new branch] gh/kurtamohler/50/orig -> origin/gh/kurtamohler/50/orig 2025-09-07T07:28:18.7990526Z * [new branch] gh/kwen2501/130/base -> origin/gh/kwen2501/130/base 2025-09-07T07:28:18.7991494Z * [new branch] gh/kwen2501/130/head -> origin/gh/kwen2501/130/head 2025-09-07T07:28:18.7992402Z * [new branch] gh/kwen2501/130/orig -> origin/gh/kwen2501/130/orig 2025-09-07T07:28:18.7993661Z * [new branch] gh/kwen2501/15/base -> origin/gh/kwen2501/15/base 2025-09-07T07:28:18.7994506Z * [new branch] gh/kwen2501/15/head -> origin/gh/kwen2501/15/head 2025-09-07T07:28:18.7995650Z * [new branch] gh/kwen2501/156/base -> origin/gh/kwen2501/156/base 2025-09-07T07:28:18.7996545Z * [new branch] gh/kwen2501/156/head -> origin/gh/kwen2501/156/head 2025-09-07T07:28:18.7997341Z * [new branch] gh/kwen2501/156/orig -> origin/gh/kwen2501/156/orig 2025-09-07T07:28:18.7998561Z * [new branch] gh/kwen2501/170/base -> origin/gh/kwen2501/170/base 2025-09-07T07:28:18.7999390Z * [new branch] gh/kwen2501/170/head -> origin/gh/kwen2501/170/head 2025-09-07T07:28:18.8000634Z * [new branch] gh/kwen2501/186/base -> origin/gh/kwen2501/186/base 2025-09-07T07:28:18.8001590Z * [new branch] gh/kwen2501/186/head -> origin/gh/kwen2501/186/head 2025-09-07T07:28:18.8002421Z * [new branch] gh/kwen2501/186/orig -> origin/gh/kwen2501/186/orig 2025-09-07T07:28:18.8003482Z * [new branch] gh/kwen2501/187/base -> origin/gh/kwen2501/187/base 2025-09-07T07:28:18.8004505Z * [new branch] gh/kwen2501/187/head -> origin/gh/kwen2501/187/head 2025-09-07T07:28:18.8005269Z * [new branch] gh/kwen2501/187/orig -> origin/gh/kwen2501/187/orig 2025-09-07T07:28:18.8006429Z * [new branch] gh/kwen2501/188/base -> origin/gh/kwen2501/188/base 2025-09-07T07:28:18.8007251Z * [new branch] gh/kwen2501/188/head -> origin/gh/kwen2501/188/head 2025-09-07T07:28:18.8008154Z * [new branch] gh/kwen2501/188/orig -> origin/gh/kwen2501/188/orig 2025-09-07T07:28:18.8009526Z * [new branch] gh/kwen2501/194/base -> origin/gh/kwen2501/194/base 2025-09-07T07:28:18.8010418Z * [new branch] gh/kwen2501/194/head -> origin/gh/kwen2501/194/head 2025-09-07T07:28:18.8011295Z * [new branch] gh/kwen2501/194/orig -> origin/gh/kwen2501/194/orig 2025-09-07T07:28:18.8012468Z * [new branch] gh/kwen2501/199/base -> origin/gh/kwen2501/199/base 2025-09-07T07:28:18.8013405Z * [new branch] gh/kwen2501/199/head -> origin/gh/kwen2501/199/head 2025-09-07T07:28:18.8014297Z * [new branch] gh/kwen2501/199/orig -> origin/gh/kwen2501/199/orig 2025-09-07T07:28:18.8015337Z * [new branch] gh/kwen2501/200/base -> origin/gh/kwen2501/200/base 2025-09-07T07:28:18.8016244Z * [new branch] gh/kwen2501/200/head -> origin/gh/kwen2501/200/head 2025-09-07T07:28:18.8017345Z * [new branch] gh/kwen2501/200/orig -> origin/gh/kwen2501/200/orig 2025-09-07T07:28:18.8018528Z * [new branch] gh/kwen2501/201/base -> origin/gh/kwen2501/201/base 2025-09-07T07:28:18.8019442Z * [new branch] gh/kwen2501/201/head -> origin/gh/kwen2501/201/head 2025-09-07T07:28:18.8020332Z * [new branch] gh/kwen2501/201/orig -> origin/gh/kwen2501/201/orig 2025-09-07T07:28:18.8021515Z * [new branch] gh/kwen2501/203/base -> origin/gh/kwen2501/203/base 2025-09-07T07:28:18.8022354Z * [new branch] gh/kwen2501/203/head -> origin/gh/kwen2501/203/head 2025-09-07T07:28:18.8023182Z * [new branch] gh/kwen2501/203/orig -> origin/gh/kwen2501/203/orig 2025-09-07T07:28:18.8024363Z * [new branch] gh/kwen2501/204/base -> origin/gh/kwen2501/204/base 2025-09-07T07:28:18.8025275Z * [new branch] gh/kwen2501/204/head -> origin/gh/kwen2501/204/head 2025-09-07T07:28:18.8026133Z * [new branch] gh/kwen2501/204/orig -> origin/gh/kwen2501/204/orig 2025-09-07T07:28:18.8027283Z * [new branch] gh/kwen2501/205/base -> origin/gh/kwen2501/205/base 2025-09-07T07:28:18.8028087Z * [new branch] gh/kwen2501/205/head -> origin/gh/kwen2501/205/head 2025-09-07T07:28:18.8028925Z * [new branch] gh/kwen2501/205/orig -> origin/gh/kwen2501/205/orig 2025-09-07T07:28:18.8030085Z * [new branch] gh/kwen2501/206/base -> origin/gh/kwen2501/206/base 2025-09-07T07:28:18.8030972Z * [new branch] gh/kwen2501/206/head -> origin/gh/kwen2501/206/head 2025-09-07T07:28:18.8031801Z * [new branch] gh/kwen2501/206/orig -> origin/gh/kwen2501/206/orig 2025-09-07T07:28:18.8032973Z * [new branch] gh/kwen2501/207/base -> origin/gh/kwen2501/207/base 2025-09-07T07:28:18.8033777Z * [new branch] gh/kwen2501/207/head -> origin/gh/kwen2501/207/head 2025-09-07T07:28:18.8034681Z * [new branch] gh/kwen2501/207/orig -> origin/gh/kwen2501/207/orig 2025-09-07T07:28:18.8035804Z * [new branch] gh/kwen2501/208/base -> origin/gh/kwen2501/208/base 2025-09-07T07:28:18.8036714Z * [new branch] gh/kwen2501/208/head -> origin/gh/kwen2501/208/head 2025-09-07T07:28:18.8037545Z * [new branch] gh/kwen2501/208/orig -> origin/gh/kwen2501/208/orig 2025-09-07T07:28:18.8039099Z * [new branch] gh/kwen2501/209/base -> origin/gh/kwen2501/209/base 2025-09-07T07:28:18.8039937Z * [new branch] gh/kwen2501/209/head -> origin/gh/kwen2501/209/head 2025-09-07T07:28:18.8040812Z * [new branch] gh/kwen2501/209/orig -> origin/gh/kwen2501/209/orig 2025-09-07T07:28:18.8042039Z * [new branch] gh/kwen2501/210/base -> origin/gh/kwen2501/210/base 2025-09-07T07:28:18.8042889Z * [new branch] gh/kwen2501/210/head -> origin/gh/kwen2501/210/head 2025-09-07T07:28:18.8043786Z * [new branch] gh/kwen2501/210/orig -> origin/gh/kwen2501/210/orig 2025-09-07T07:28:18.8044927Z * [new branch] gh/kwen2501/211/base -> origin/gh/kwen2501/211/base 2025-09-07T07:28:18.8045827Z * [new branch] gh/kwen2501/211/head -> origin/gh/kwen2501/211/head 2025-09-07T07:28:18.8047004Z * [new branch] gh/kwen2501/212/base -> origin/gh/kwen2501/212/base 2025-09-07T07:28:18.8047977Z * [new branch] gh/kwen2501/212/head -> origin/gh/kwen2501/212/head 2025-09-07T07:28:18.8048883Z * [new branch] gh/kwen2501/212/orig -> origin/gh/kwen2501/212/orig 2025-09-07T07:28:18.8050038Z * [new branch] gh/kwen2501/213/base -> origin/gh/kwen2501/213/base 2025-09-07T07:28:18.8050887Z * [new branch] gh/kwen2501/213/head -> origin/gh/kwen2501/213/head 2025-09-07T07:28:18.8051726Z * [new branch] gh/kwen2501/213/orig -> origin/gh/kwen2501/213/orig 2025-09-07T07:28:18.8052913Z * [new branch] gh/kwen2501/214/base -> origin/gh/kwen2501/214/base 2025-09-07T07:28:18.8053752Z * [new branch] gh/kwen2501/214/head -> origin/gh/kwen2501/214/head 2025-09-07T07:28:18.8054603Z * [new branch] gh/kwen2501/214/orig -> origin/gh/kwen2501/214/orig 2025-09-07T07:28:18.8055844Z * [new branch] gh/kwen2501/215/base -> origin/gh/kwen2501/215/base 2025-09-07T07:28:18.8056743Z * [new branch] gh/kwen2501/215/head -> origin/gh/kwen2501/215/head 2025-09-07T07:28:18.8057577Z * [new branch] gh/kwen2501/215/orig -> origin/gh/kwen2501/215/orig 2025-09-07T07:28:18.8058720Z * [new branch] gh/kwen2501/216/base -> origin/gh/kwen2501/216/base 2025-09-07T07:28:18.8059553Z * [new branch] gh/kwen2501/216/head -> origin/gh/kwen2501/216/head 2025-09-07T07:28:18.8060449Z * [new branch] gh/kwen2501/216/orig -> origin/gh/kwen2501/216/orig 2025-09-07T07:28:18.8061578Z * [new branch] gh/kwen2501/217/base -> origin/gh/kwen2501/217/base 2025-09-07T07:28:18.8062403Z * [new branch] gh/kwen2501/217/head -> origin/gh/kwen2501/217/head 2025-09-07T07:28:18.8063245Z * [new branch] gh/kwen2501/217/orig -> origin/gh/kwen2501/217/orig 2025-09-07T07:28:18.8064437Z * [new branch] gh/kwen2501/218/base -> origin/gh/kwen2501/218/base 2025-09-07T07:28:18.8065382Z * [new branch] gh/kwen2501/218/head -> origin/gh/kwen2501/218/head 2025-09-07T07:28:18.8066205Z * [new branch] gh/kwen2501/218/orig -> origin/gh/kwen2501/218/orig 2025-09-07T07:28:18.8067381Z * [new branch] gh/kwen2501/219/base -> origin/gh/kwen2501/219/base 2025-09-07T07:28:18.8068268Z * [new branch] gh/kwen2501/219/head -> origin/gh/kwen2501/219/head 2025-09-07T07:28:18.8069328Z * [new branch] gh/kwen2501/219/orig -> origin/gh/kwen2501/219/orig 2025-09-07T07:28:18.8070585Z * [new branch] gh/kwen2501/220/base -> origin/gh/kwen2501/220/base 2025-09-07T07:28:18.8071447Z * [new branch] gh/kwen2501/220/head -> origin/gh/kwen2501/220/head 2025-09-07T07:28:18.8072344Z * [new branch] gh/kwen2501/220/orig -> origin/gh/kwen2501/220/orig 2025-09-07T07:28:18.8073689Z * [new branch] gh/kwen2501/221/base -> origin/gh/kwen2501/221/base 2025-09-07T07:28:18.8074372Z * [new branch] gh/kwen2501/221/head -> origin/gh/kwen2501/221/head 2025-09-07T07:28:18.8075211Z * [new branch] gh/kwen2501/221/orig -> origin/gh/kwen2501/221/orig 2025-09-07T07:28:18.8076414Z * [new branch] gh/kwen2501/222/base -> origin/gh/kwen2501/222/base 2025-09-07T07:28:18.8077277Z * [new branch] gh/kwen2501/222/head -> origin/gh/kwen2501/222/head 2025-09-07T07:28:18.8078152Z * [new branch] gh/kwen2501/222/orig -> origin/gh/kwen2501/222/orig 2025-09-07T07:28:18.8079294Z * [new branch] gh/kwen2501/223/base -> origin/gh/kwen2501/223/base 2025-09-07T07:28:18.8080109Z * [new branch] gh/kwen2501/223/head -> origin/gh/kwen2501/223/head 2025-09-07T07:28:18.8080980Z * [new branch] gh/kwen2501/223/orig -> origin/gh/kwen2501/223/orig 2025-09-07T07:28:18.8082165Z * [new branch] gh/kwen2501/224/base -> origin/gh/kwen2501/224/base 2025-09-07T07:28:18.8083029Z * [new branch] gh/kwen2501/224/head -> origin/gh/kwen2501/224/head 2025-09-07T07:28:18.8083931Z * [new branch] gh/kwen2501/224/orig -> origin/gh/kwen2501/224/orig 2025-09-07T07:28:18.8085134Z * [new branch] gh/kwen2501/225/base -> origin/gh/kwen2501/225/base 2025-09-07T07:28:18.8085996Z * [new branch] gh/kwen2501/225/head -> origin/gh/kwen2501/225/head 2025-09-07T07:28:18.8086832Z * [new branch] gh/kwen2501/225/orig -> origin/gh/kwen2501/225/orig 2025-09-07T07:28:18.8087961Z * [new branch] gh/kwen2501/226/base -> origin/gh/kwen2501/226/base 2025-09-07T07:28:18.8088838Z * [new branch] gh/kwen2501/226/head -> origin/gh/kwen2501/226/head 2025-09-07T07:28:18.8089740Z * [new branch] gh/kwen2501/226/orig -> origin/gh/kwen2501/226/orig 2025-09-07T07:28:18.8090918Z * [new branch] gh/kwen2501/227/base -> origin/gh/kwen2501/227/base 2025-09-07T07:28:18.8091730Z * [new branch] gh/kwen2501/227/head -> origin/gh/kwen2501/227/head 2025-09-07T07:28:18.8092584Z * [new branch] gh/kwen2501/227/orig -> origin/gh/kwen2501/227/orig 2025-09-07T07:28:18.8093821Z * [new branch] gh/kwen2501/228/base -> origin/gh/kwen2501/228/base 2025-09-07T07:28:18.8094649Z * [new branch] gh/kwen2501/228/head -> origin/gh/kwen2501/228/head 2025-09-07T07:28:18.8095600Z * [new branch] gh/kwen2501/228/orig -> origin/gh/kwen2501/228/orig 2025-09-07T07:28:18.8096779Z * [new branch] gh/kwen2501/229/base -> origin/gh/kwen2501/229/base 2025-09-07T07:28:18.8097648Z * [new branch] gh/kwen2501/229/head -> origin/gh/kwen2501/229/head 2025-09-07T07:28:18.8098459Z * [new branch] gh/kwen2501/229/orig -> origin/gh/kwen2501/229/orig 2025-09-07T07:28:18.8099677Z * [new branch] gh/kwen2501/230/base -> origin/gh/kwen2501/230/base 2025-09-07T07:28:18.8100519Z * [new branch] gh/kwen2501/230/head -> origin/gh/kwen2501/230/head 2025-09-07T07:28:18.8101370Z * [new branch] gh/kwen2501/230/orig -> origin/gh/kwen2501/230/orig 2025-09-07T07:28:18.8102609Z * [new branch] gh/kwen2501/231/base -> origin/gh/kwen2501/231/base 2025-09-07T07:28:18.8103458Z * [new branch] gh/kwen2501/231/head -> origin/gh/kwen2501/231/head 2025-09-07T07:28:18.8104267Z * [new branch] gh/kwen2501/231/orig -> origin/gh/kwen2501/231/orig 2025-09-07T07:28:18.8105544Z * [new branch] gh/kwen2501/232/base -> origin/gh/kwen2501/232/base 2025-09-07T07:28:18.8106337Z * [new branch] gh/kwen2501/232/head -> origin/gh/kwen2501/232/head 2025-09-07T07:28:18.8107219Z * [new branch] gh/kwen2501/232/orig -> origin/gh/kwen2501/232/orig 2025-09-07T07:28:18.8108972Z * [new branch] gh/laithsakka/156/base -> origin/gh/laithsakka/156/base 2025-09-07T07:28:18.8109723Z * [new branch] gh/laithsakka/156/head -> origin/gh/laithsakka/156/head 2025-09-07T07:28:18.8110558Z * [new branch] gh/laithsakka/156/orig -> origin/gh/laithsakka/156/orig 2025-09-07T07:28:18.8111925Z * [new branch] gh/laithsakka/160/base -> origin/gh/laithsakka/160/base 2025-09-07T07:28:18.8112723Z * [new branch] gh/laithsakka/160/head -> origin/gh/laithsakka/160/head 2025-09-07T07:28:18.8113568Z * [new branch] gh/laithsakka/160/orig -> origin/gh/laithsakka/160/orig 2025-09-07T07:28:18.8114770Z * [new branch] gh/laithsakka/178/base -> origin/gh/laithsakka/178/base 2025-09-07T07:28:18.8115664Z * [new branch] gh/laithsakka/178/head -> origin/gh/laithsakka/178/head 2025-09-07T07:28:18.8116516Z * [new branch] gh/laithsakka/178/orig -> origin/gh/laithsakka/178/orig 2025-09-07T07:28:18.8117689Z * [new branch] gh/laithsakka/191/base -> origin/gh/laithsakka/191/base 2025-09-07T07:28:18.8118649Z * [new branch] gh/laithsakka/191/head -> origin/gh/laithsakka/191/head 2025-09-07T07:28:18.8119467Z * [new branch] gh/laithsakka/191/orig -> origin/gh/laithsakka/191/orig 2025-09-07T07:28:18.8120663Z * [new branch] gh/laithsakka/237/base -> origin/gh/laithsakka/237/base 2025-09-07T07:28:18.8121465Z * [new branch] gh/laithsakka/237/head -> origin/gh/laithsakka/237/head 2025-09-07T07:28:18.8122342Z * [new branch] gh/laithsakka/237/orig -> origin/gh/laithsakka/237/orig 2025-09-07T07:28:18.8123460Z * [new branch] gh/laithsakka/249/base -> origin/gh/laithsakka/249/base 2025-09-07T07:28:18.8124354Z * [new branch] gh/laithsakka/249/head -> origin/gh/laithsakka/249/head 2025-09-07T07:28:18.8125158Z * [new branch] gh/laithsakka/249/orig -> origin/gh/laithsakka/249/orig 2025-09-07T07:28:18.8126335Z * [new branch] gh/laithsakka/251/base -> origin/gh/laithsakka/251/base 2025-09-07T07:28:18.8127177Z * [new branch] gh/laithsakka/251/head -> origin/gh/laithsakka/251/head 2025-09-07T07:28:18.8128044Z * [new branch] gh/laithsakka/251/orig -> origin/gh/laithsakka/251/orig 2025-09-07T07:28:18.8129220Z * [new branch] gh/laithsakka/254/base -> origin/gh/laithsakka/254/base 2025-09-07T07:28:18.8130122Z * [new branch] gh/laithsakka/254/head -> origin/gh/laithsakka/254/head 2025-09-07T07:28:18.8130915Z * [new branch] gh/laithsakka/254/orig -> origin/gh/laithsakka/254/orig 2025-09-07T07:28:18.8132182Z * [new branch] gh/laithsakka/255/base -> origin/gh/laithsakka/255/base 2025-09-07T07:28:18.8133051Z * [new branch] gh/laithsakka/255/head -> origin/gh/laithsakka/255/head 2025-09-07T07:28:18.8133801Z * [new branch] gh/laithsakka/255/orig -> origin/gh/laithsakka/255/orig 2025-09-07T07:28:18.8134892Z * [new branch] gh/laithsakka/256/base -> origin/gh/laithsakka/256/base 2025-09-07T07:28:18.8135840Z * [new branch] gh/laithsakka/256/head -> origin/gh/laithsakka/256/head 2025-09-07T07:28:18.8136636Z * [new branch] gh/laithsakka/256/orig -> origin/gh/laithsakka/256/orig 2025-09-07T07:28:18.8137805Z * [new branch] gh/laithsakka/257/base -> origin/gh/laithsakka/257/base 2025-09-07T07:28:18.8138643Z * [new branch] gh/laithsakka/257/head -> origin/gh/laithsakka/257/head 2025-09-07T07:28:18.8139494Z * [new branch] gh/laithsakka/257/orig -> origin/gh/laithsakka/257/orig 2025-09-07T07:28:18.8140708Z * [new branch] gh/laithsakka/258/base -> origin/gh/laithsakka/258/base 2025-09-07T07:28:18.8141660Z * [new branch] gh/laithsakka/258/head -> origin/gh/laithsakka/258/head 2025-09-07T07:28:18.8142395Z * [new branch] gh/laithsakka/258/orig -> origin/gh/laithsakka/258/orig 2025-09-07T07:28:18.8143845Z * [new branch] gh/laithsakka/259/base -> origin/gh/laithsakka/259/base 2025-09-07T07:28:18.8144765Z * [new branch] gh/laithsakka/259/head -> origin/gh/laithsakka/259/head 2025-09-07T07:28:18.8145615Z * [new branch] gh/laithsakka/259/orig -> origin/gh/laithsakka/259/orig 2025-09-07T07:28:18.8146713Z * [new branch] gh/laithsakka/260/base -> origin/gh/laithsakka/260/base 2025-09-07T07:28:18.8147609Z * [new branch] gh/laithsakka/260/head -> origin/gh/laithsakka/260/head 2025-09-07T07:28:18.8148491Z * [new branch] gh/laithsakka/260/orig -> origin/gh/laithsakka/260/orig 2025-09-07T07:28:18.8149638Z * [new branch] gh/laithsakka/261/base -> origin/gh/laithsakka/261/base 2025-09-07T07:28:18.8150505Z * [new branch] gh/laithsakka/261/head -> origin/gh/laithsakka/261/head 2025-09-07T07:28:18.8159001Z * [new branch] gh/laithsakka/261/orig -> origin/gh/laithsakka/261/orig 2025-09-07T07:28:18.8159276Z * [new branch] gh/laithsakka/262/base -> origin/gh/laithsakka/262/base 2025-09-07T07:28:18.8159481Z * [new branch] gh/laithsakka/262/head -> origin/gh/laithsakka/262/head 2025-09-07T07:28:18.8159678Z * [new branch] gh/laithsakka/262/orig -> origin/gh/laithsakka/262/orig 2025-09-07T07:28:18.8159879Z * [new branch] gh/laithsakka/263/base -> origin/gh/laithsakka/263/base 2025-09-07T07:28:18.8160088Z * [new branch] gh/laithsakka/263/head -> origin/gh/laithsakka/263/head 2025-09-07T07:28:18.8160282Z * [new branch] gh/laithsakka/263/orig -> origin/gh/laithsakka/263/orig 2025-09-07T07:28:18.8160490Z * [new branch] gh/laithsakka/264/base -> origin/gh/laithsakka/264/base 2025-09-07T07:28:18.8160685Z * [new branch] gh/laithsakka/264/head -> origin/gh/laithsakka/264/head 2025-09-07T07:28:18.8161213Z * [new branch] gh/laithsakka/264/orig -> origin/gh/laithsakka/264/orig 2025-09-07T07:28:18.8162528Z * [new branch] gh/laithsakka/265/base -> origin/gh/laithsakka/265/base 2025-09-07T07:28:18.8163344Z * [new branch] gh/laithsakka/265/head -> origin/gh/laithsakka/265/head 2025-09-07T07:28:18.8164178Z * [new branch] gh/laithsakka/265/orig -> origin/gh/laithsakka/265/orig 2025-09-07T07:28:18.8165413Z * [new branch] gh/laithsakka/266/base -> origin/gh/laithsakka/266/base 2025-09-07T07:28:18.8166389Z * [new branch] gh/laithsakka/266/head -> origin/gh/laithsakka/266/head 2025-09-07T07:28:18.8167241Z * [new branch] gh/laithsakka/266/orig -> origin/gh/laithsakka/266/orig 2025-09-07T07:28:18.8168402Z * [new branch] gh/laithsakka/267/base -> origin/gh/laithsakka/267/base 2025-09-07T07:28:18.8169238Z * [new branch] gh/laithsakka/267/head -> origin/gh/laithsakka/267/head 2025-09-07T07:28:18.8170070Z * [new branch] gh/laithsakka/267/orig -> origin/gh/laithsakka/267/orig 2025-09-07T07:28:18.8171313Z * [new branch] gh/laithsakka/268/base -> origin/gh/laithsakka/268/base 2025-09-07T07:28:18.8172148Z * [new branch] gh/laithsakka/268/head -> origin/gh/laithsakka/268/head 2025-09-07T07:28:18.8172982Z * [new branch] gh/laithsakka/268/orig -> origin/gh/laithsakka/268/orig 2025-09-07T07:28:18.8174463Z * [new branch] gh/laithsakka/28/base -> origin/gh/laithsakka/28/base 2025-09-07T07:28:18.8175513Z * [new branch] gh/laithsakka/29/base -> origin/gh/laithsakka/29/base 2025-09-07T07:28:18.8176745Z * [new branch] gh/laithsakka/30/base -> origin/gh/laithsakka/30/base 2025-09-07T07:28:18.8177565Z * [new branch] gh/laithsakka/30/head -> origin/gh/laithsakka/30/head 2025-09-07T07:28:18.8178663Z * [new branch] gh/laithsakka/31/base -> origin/gh/laithsakka/31/base 2025-09-07T07:28:18.8179475Z * [new branch] gh/laithsakka/31/head -> origin/gh/laithsakka/31/head 2025-09-07T07:28:18.8180549Z * [new branch] gh/laithsakka/32/base -> origin/gh/laithsakka/32/base 2025-09-07T07:28:18.8181360Z * [new branch] gh/laithsakka/32/head -> origin/gh/laithsakka/32/head 2025-09-07T07:28:18.8184637Z * [new branch] gh/lucaskabela/1/base -> origin/gh/lucaskabela/1/base 2025-09-07T07:28:18.8185447Z * [new branch] gh/lucaskabela/1/head -> origin/gh/lucaskabela/1/head 2025-09-07T07:28:18.8186718Z * [new branch] gh/lucaskabela/10/base -> origin/gh/lucaskabela/10/base 2025-09-07T07:28:18.8187557Z * [new branch] gh/lucaskabela/10/head -> origin/gh/lucaskabela/10/head 2025-09-07T07:28:18.8188380Z * [new branch] gh/lucaskabela/10/orig -> origin/gh/lucaskabela/10/orig 2025-09-07T07:28:18.8189579Z * [new branch] gh/lucaskabela/11/base -> origin/gh/lucaskabela/11/base 2025-09-07T07:28:18.8190455Z * [new branch] gh/lucaskabela/11/head -> origin/gh/lucaskabela/11/head 2025-09-07T07:28:18.8191314Z * [new branch] gh/lucaskabela/11/orig -> origin/gh/lucaskabela/11/orig 2025-09-07T07:28:18.8192333Z * [new branch] gh/lucaskabela/12/base -> origin/gh/lucaskabela/12/base 2025-09-07T07:28:18.8193200Z * [new branch] gh/lucaskabela/12/head -> origin/gh/lucaskabela/12/head 2025-09-07T07:28:18.8194054Z * [new branch] gh/lucaskabela/12/orig -> origin/gh/lucaskabela/12/orig 2025-09-07T07:28:18.8195081Z * [new branch] gh/lucaskabela/13/base -> origin/gh/lucaskabela/13/base 2025-09-07T07:28:18.8195912Z * [new branch] gh/lucaskabela/13/head -> origin/gh/lucaskabela/13/head 2025-09-07T07:28:18.8196763Z * [new branch] gh/lucaskabela/13/orig -> origin/gh/lucaskabela/13/orig 2025-09-07T07:28:18.8197850Z * [new branch] gh/lucaskabela/14/base -> origin/gh/lucaskabela/14/base 2025-09-07T07:28:18.8198705Z * [new branch] gh/lucaskabela/14/head -> origin/gh/lucaskabela/14/head 2025-09-07T07:28:18.8199513Z * [new branch] gh/lucaskabela/14/orig -> origin/gh/lucaskabela/14/orig 2025-09-07T07:28:18.8200665Z * [new branch] gh/lucaskabela/15/base -> origin/gh/lucaskabela/15/base 2025-09-07T07:28:18.8201484Z * [new branch] gh/lucaskabela/15/head -> origin/gh/lucaskabela/15/head 2025-09-07T07:28:18.8202341Z * [new branch] gh/lucaskabela/15/orig -> origin/gh/lucaskabela/15/orig 2025-09-07T07:28:18.8203441Z * [new branch] gh/lucaskabela/16/base -> origin/gh/lucaskabela/16/base 2025-09-07T07:28:18.8204243Z * [new branch] gh/lucaskabela/16/head -> origin/gh/lucaskabela/16/head 2025-09-07T07:28:18.8205079Z * [new branch] gh/lucaskabela/16/orig -> origin/gh/lucaskabela/16/orig 2025-09-07T07:28:18.8206137Z * [new branch] gh/lucaskabela/17/base -> origin/gh/lucaskabela/17/base 2025-09-07T07:28:18.8207011Z * [new branch] gh/lucaskabela/17/head -> origin/gh/lucaskabela/17/head 2025-09-07T07:28:18.8207856Z * [new branch] gh/lucaskabela/17/orig -> origin/gh/lucaskabela/17/orig 2025-09-07T07:28:18.8208972Z * [new branch] gh/lucaskabela/2/base -> origin/gh/lucaskabela/2/base 2025-09-07T07:28:18.8209908Z * [new branch] gh/lucaskabela/2/head -> origin/gh/lucaskabela/2/head 2025-09-07T07:28:18.8210729Z * [new branch] gh/lucaskabela/2/orig -> origin/gh/lucaskabela/2/orig 2025-09-07T07:28:18.8212106Z * [new branch] gh/lucaskabela/3/base -> origin/gh/lucaskabela/3/base 2025-09-07T07:28:18.8212819Z * [new branch] gh/lucaskabela/3/head -> origin/gh/lucaskabela/3/head 2025-09-07T07:28:18.8213670Z * [new branch] gh/lucaskabela/3/orig -> origin/gh/lucaskabela/3/orig 2025-09-07T07:28:18.8214776Z * [new branch] gh/lucaskabela/4/base -> origin/gh/lucaskabela/4/base 2025-09-07T07:28:18.8215615Z * [new branch] gh/lucaskabela/4/head -> origin/gh/lucaskabela/4/head 2025-09-07T07:28:18.8216532Z * [new branch] gh/lucaskabela/4/orig -> origin/gh/lucaskabela/4/orig 2025-09-07T07:28:18.8217768Z * [new branch] gh/lucaskabela/5/base -> origin/gh/lucaskabela/5/base 2025-09-07T07:28:18.8218566Z * [new branch] gh/lucaskabela/5/head -> origin/gh/lucaskabela/5/head 2025-09-07T07:28:18.8219422Z * [new branch] gh/lucaskabela/5/orig -> origin/gh/lucaskabela/5/orig 2025-09-07T07:28:18.8220522Z * [new branch] gh/lucaskabela/6/base -> origin/gh/lucaskabela/6/base 2025-09-07T07:28:18.8221338Z * [new branch] gh/lucaskabela/6/head -> origin/gh/lucaskabela/6/head 2025-09-07T07:28:18.8222179Z * [new branch] gh/lucaskabela/6/orig -> origin/gh/lucaskabela/6/orig 2025-09-07T07:28:18.8223466Z * [new branch] gh/lucaskabela/7/base -> origin/gh/lucaskabela/7/base 2025-09-07T07:28:18.8224250Z * [new branch] gh/lucaskabela/7/head -> origin/gh/lucaskabela/7/head 2025-09-07T07:28:18.8225077Z * [new branch] gh/lucaskabela/7/orig -> origin/gh/lucaskabela/7/orig 2025-09-07T07:28:18.8226198Z * [new branch] gh/lucaskabela/8/base -> origin/gh/lucaskabela/8/base 2025-09-07T07:28:18.8227117Z * [new branch] gh/lucaskabela/8/head -> origin/gh/lucaskabela/8/head 2025-09-07T07:28:18.8227973Z * [new branch] gh/lucaskabela/8/orig -> origin/gh/lucaskabela/8/orig 2025-09-07T07:28:18.8229102Z * [new branch] gh/lucaskabela/9/base -> origin/gh/lucaskabela/9/base 2025-09-07T07:28:18.8229980Z * [new branch] gh/lucaskabela/9/head -> origin/gh/lucaskabela/9/head 2025-09-07T07:28:18.8230795Z * [new branch] gh/lucaskabela/9/orig -> origin/gh/lucaskabela/9/orig 2025-09-07T07:28:18.8232203Z * [new branch] gh/lw/3/base -> origin/gh/lw/3/base 2025-09-07T07:28:18.8233102Z * [new branch] gh/lw/3/head -> origin/gh/lw/3/head 2025-09-07T07:28:18.8233913Z * [new branch] gh/lw/3/orig -> origin/gh/lw/3/orig 2025-09-07T07:28:18.8235424Z * [new branch] gh/malfet/14/base -> origin/gh/malfet/14/base 2025-09-07T07:28:18.8236781Z * [new branch] gh/malfet/330/base -> origin/gh/malfet/330/base 2025-09-07T07:28:18.8237495Z * [new branch] gh/malfet/330/head -> origin/gh/malfet/330/head 2025-09-07T07:28:18.8238330Z * [new branch] gh/malfet/330/orig -> origin/gh/malfet/330/orig 2025-09-07T07:28:18.8239487Z * [new branch] gh/malfet/396/base -> origin/gh/malfet/396/base 2025-09-07T07:28:18.8240336Z * [new branch] gh/malfet/396/head -> origin/gh/malfet/396/head 2025-09-07T07:28:18.8241358Z * [new branch] gh/malfet/396/orig -> origin/gh/malfet/396/orig 2025-09-07T07:28:18.8242377Z * [new branch] gh/malfet/397/base -> origin/gh/malfet/397/base 2025-09-07T07:28:18.8243212Z * [new branch] gh/malfet/397/head -> origin/gh/malfet/397/head 2025-09-07T07:28:18.8244073Z * [new branch] gh/malfet/397/orig -> origin/gh/malfet/397/orig 2025-09-07T07:28:18.8247065Z * [new branch] gh/malfet/398/base -> origin/gh/malfet/398/base 2025-09-07T07:28:18.8247581Z * [new branch] gh/malfet/398/head -> origin/gh/malfet/398/head 2025-09-07T07:28:18.8247765Z * [new branch] gh/malfet/398/orig -> origin/gh/malfet/398/orig 2025-09-07T07:28:18.8248472Z * [new branch] gh/malfet/399/base -> origin/gh/malfet/399/base 2025-09-07T07:28:18.8249213Z * [new branch] gh/malfet/399/head -> origin/gh/malfet/399/head 2025-09-07T07:28:18.8249930Z * [new branch] gh/malfet/399/orig -> origin/gh/malfet/399/orig 2025-09-07T07:28:18.8251096Z * [new branch] gh/malfet/414/base -> origin/gh/malfet/414/base 2025-09-07T07:28:18.8251948Z * [new branch] gh/malfet/414/head -> origin/gh/malfet/414/head 2025-09-07T07:28:18.8252814Z * [new branch] gh/malfet/414/orig -> origin/gh/malfet/414/orig 2025-09-07T07:28:18.8253985Z * [new branch] gh/malfet/417/base -> origin/gh/malfet/417/base 2025-09-07T07:28:18.8254826Z * [new branch] gh/malfet/417/head -> origin/gh/malfet/417/head 2025-09-07T07:28:18.8255668Z * [new branch] gh/malfet/417/orig -> origin/gh/malfet/417/orig 2025-09-07T07:28:18.8256838Z * [new branch] gh/malfet/418/base -> origin/gh/malfet/418/base 2025-09-07T07:28:18.8257700Z * [new branch] gh/malfet/418/head -> origin/gh/malfet/418/head 2025-09-07T07:28:18.8258496Z * [new branch] gh/malfet/418/orig -> origin/gh/malfet/418/orig 2025-09-07T07:28:18.8259758Z * [new branch] gh/malfet/475/base -> origin/gh/malfet/475/base 2025-09-07T07:28:18.8260660Z * [new branch] gh/malfet/475/head -> origin/gh/malfet/475/head 2025-09-07T07:28:18.8261508Z * [new branch] gh/malfet/475/orig -> origin/gh/malfet/475/orig 2025-09-07T07:28:18.8262688Z * [new branch] gh/malfet/476/base -> origin/gh/malfet/476/base 2025-09-07T07:28:18.8263574Z * [new branch] gh/malfet/476/head -> origin/gh/malfet/476/head 2025-09-07T07:28:18.8264371Z * [new branch] gh/malfet/476/orig -> origin/gh/malfet/476/orig 2025-09-07T07:28:18.8265638Z * [new branch] gh/malfet/477/base -> origin/gh/malfet/477/base 2025-09-07T07:28:18.8266467Z * [new branch] gh/malfet/477/head -> origin/gh/malfet/477/head 2025-09-07T07:28:18.8267332Z * [new branch] gh/malfet/477/orig -> origin/gh/malfet/477/orig 2025-09-07T07:28:18.8268479Z * [new branch] gh/malfet/478/base -> origin/gh/malfet/478/base 2025-09-07T07:28:18.8269340Z * [new branch] gh/malfet/478/head -> origin/gh/malfet/478/head 2025-09-07T07:28:18.8270226Z * [new branch] gh/malfet/478/orig -> origin/gh/malfet/478/orig 2025-09-07T07:28:18.8271550Z * [new branch] gh/malfet/479/base -> origin/gh/malfet/479/base 2025-09-07T07:28:18.8272460Z * [new branch] gh/malfet/479/head -> origin/gh/malfet/479/head 2025-09-07T07:28:18.8273466Z * [new branch] gh/malfet/479/orig -> origin/gh/malfet/479/orig 2025-09-07T07:28:18.8274648Z * [new branch] gh/malfet/480/base -> origin/gh/malfet/480/base 2025-09-07T07:28:18.8275522Z * [new branch] gh/malfet/480/head -> origin/gh/malfet/480/head 2025-09-07T07:28:18.8276363Z * [new branch] gh/malfet/480/orig -> origin/gh/malfet/480/orig 2025-09-07T07:28:18.8277600Z * [new branch] gh/malfet/481/base -> origin/gh/malfet/481/base 2025-09-07T07:28:18.8278382Z * [new branch] gh/malfet/481/head -> origin/gh/malfet/481/head 2025-09-07T07:28:18.8279250Z * [new branch] gh/malfet/481/orig -> origin/gh/malfet/481/orig 2025-09-07T07:28:18.8280594Z * [new branch] gh/malfet/482/base -> origin/gh/malfet/482/base 2025-09-07T07:28:18.8281691Z * [new branch] gh/malfet/482/head -> origin/gh/malfet/482/head 2025-09-07T07:28:18.8282400Z * [new branch] gh/malfet/482/orig -> origin/gh/malfet/482/orig 2025-09-07T07:28:18.8283918Z * [new branch] gh/malfet/483/base -> origin/gh/malfet/483/base 2025-09-07T07:28:18.8284806Z * [new branch] gh/malfet/483/head -> origin/gh/malfet/483/head 2025-09-07T07:28:18.8285640Z * [new branch] gh/malfet/483/orig -> origin/gh/malfet/483/orig 2025-09-07T07:28:18.8286813Z * [new branch] gh/malfet/484/base -> origin/gh/malfet/484/base 2025-09-07T07:28:18.8287679Z * [new branch] gh/malfet/484/head -> origin/gh/malfet/484/head 2025-09-07T07:28:18.8288617Z * [new branch] gh/malfet/484/orig -> origin/gh/malfet/484/orig 2025-09-07T07:28:18.8289864Z * [new branch] gh/malfet/485/base -> origin/gh/malfet/485/base 2025-09-07T07:28:18.8290709Z * [new branch] gh/malfet/485/head -> origin/gh/malfet/485/head 2025-09-07T07:28:18.8291648Z * [new branch] gh/malfet/485/orig -> origin/gh/malfet/485/orig 2025-09-07T07:28:18.8292871Z * [new branch] gh/malfet/486/base -> origin/gh/malfet/486/base 2025-09-07T07:28:18.8293724Z * [new branch] gh/malfet/486/head -> origin/gh/malfet/486/head 2025-09-07T07:28:18.8294558Z * [new branch] gh/malfet/486/orig -> origin/gh/malfet/486/orig 2025-09-07T07:28:18.8295900Z * [new branch] gh/malfet/487/base -> origin/gh/malfet/487/base 2025-09-07T07:28:18.8296792Z * [new branch] gh/malfet/487/head -> origin/gh/malfet/487/head 2025-09-07T07:28:18.8297659Z * [new branch] gh/malfet/487/orig -> origin/gh/malfet/487/orig 2025-09-07T07:28:18.8299084Z * [new branch] gh/malfet/488/base -> origin/gh/malfet/488/base 2025-09-07T07:28:18.8299921Z * [new branch] gh/malfet/488/head -> origin/gh/malfet/488/head 2025-09-07T07:28:18.8300815Z * [new branch] gh/malfet/488/orig -> origin/gh/malfet/488/orig 2025-09-07T07:28:18.8301951Z * [new branch] gh/malfet/489/base -> origin/gh/malfet/489/base 2025-09-07T07:28:18.8302868Z * [new branch] gh/malfet/489/head -> origin/gh/malfet/489/head 2025-09-07T07:28:18.8303945Z * [new branch] gh/malfet/489/orig -> origin/gh/malfet/489/orig 2025-09-07T07:28:18.8305180Z * [new branch] gh/malfet/490/base -> origin/gh/malfet/490/base 2025-09-07T07:28:18.8306006Z * [new branch] gh/malfet/490/head -> origin/gh/malfet/490/head 2025-09-07T07:28:18.8306895Z * [new branch] gh/malfet/490/orig -> origin/gh/malfet/490/orig 2025-09-07T07:28:18.8308211Z * [new branch] gh/malfet/491/base -> origin/gh/malfet/491/base 2025-09-07T07:28:18.8309073Z * [new branch] gh/malfet/491/head -> origin/gh/malfet/491/head 2025-09-07T07:28:18.8309942Z * [new branch] gh/malfet/491/orig -> origin/gh/malfet/491/orig 2025-09-07T07:28:18.8311035Z * [new branch] gh/malfet/492/base -> origin/gh/malfet/492/base 2025-09-07T07:28:18.8312115Z * [new branch] gh/malfet/492/head -> origin/gh/malfet/492/head 2025-09-07T07:28:18.8313005Z * [new branch] gh/malfet/492/orig -> origin/gh/malfet/492/orig 2025-09-07T07:28:18.8314260Z * [new branch] gh/malfet/493/base -> origin/gh/malfet/493/base 2025-09-07T07:28:18.8315115Z * [new branch] gh/malfet/493/head -> origin/gh/malfet/493/head 2025-09-07T07:28:18.8315936Z * [new branch] gh/malfet/493/orig -> origin/gh/malfet/493/orig 2025-09-07T07:28:18.8317192Z * [new branch] gh/malfet/494/base -> origin/gh/malfet/494/base 2025-09-07T07:28:18.8318025Z * [new branch] gh/malfet/494/head -> origin/gh/malfet/494/head 2025-09-07T07:28:18.8318930Z * [new branch] gh/malfet/494/orig -> origin/gh/malfet/494/orig 2025-09-07T07:28:18.8320040Z * [new branch] gh/malfet/495/base -> origin/gh/malfet/495/base 2025-09-07T07:28:18.8320934Z * [new branch] gh/malfet/495/head -> origin/gh/malfet/495/head 2025-09-07T07:28:18.8321770Z * [new branch] gh/malfet/495/orig -> origin/gh/malfet/495/orig 2025-09-07T07:28:18.8322991Z * [new branch] gh/malfet/496/base -> origin/gh/malfet/496/base 2025-09-07T07:28:18.8323841Z * [new branch] gh/malfet/496/head -> origin/gh/malfet/496/head 2025-09-07T07:28:18.8324669Z * [new branch] gh/malfet/496/orig -> origin/gh/malfet/496/orig 2025-09-07T07:28:18.8325890Z * [new branch] gh/malfet/497/base -> origin/gh/malfet/497/base 2025-09-07T07:28:18.8326716Z * [new branch] gh/malfet/497/head -> origin/gh/malfet/497/head 2025-09-07T07:28:18.8327669Z * [new branch] gh/malfet/497/orig -> origin/gh/malfet/497/orig 2025-09-07T07:28:18.8328940Z * [new branch] gh/malfet/498/base -> origin/gh/malfet/498/base 2025-09-07T07:28:18.8329802Z * [new branch] gh/malfet/498/head -> origin/gh/malfet/498/head 2025-09-07T07:28:18.8330636Z * [new branch] gh/malfet/498/orig -> origin/gh/malfet/498/orig 2025-09-07T07:28:18.8331798Z * [new branch] gh/malfet/499/base -> origin/gh/malfet/499/base 2025-09-07T07:28:18.8332630Z * [new branch] gh/malfet/499/head -> origin/gh/malfet/499/head 2025-09-07T07:28:18.8333476Z * [new branch] gh/malfet/499/orig -> origin/gh/malfet/499/orig 2025-09-07T07:28:18.8334699Z * [new branch] gh/malfet/500/base -> origin/gh/malfet/500/base 2025-09-07T07:28:18.8335523Z * [new branch] gh/malfet/500/head -> origin/gh/malfet/500/head 2025-09-07T07:28:18.8336574Z * [new branch] gh/malfet/500/orig -> origin/gh/malfet/500/orig 2025-09-07T07:28:18.8337855Z * [new branch] gh/malfet/501/base -> origin/gh/malfet/501/base 2025-09-07T07:28:18.8338740Z * [new branch] gh/malfet/501/head -> origin/gh/malfet/501/head 2025-09-07T07:28:18.8339579Z * [new branch] gh/malfet/501/orig -> origin/gh/malfet/501/orig 2025-09-07T07:28:18.8340854Z * [new branch] gh/malfet/502/base -> origin/gh/malfet/502/base 2025-09-07T07:28:18.8341696Z * [new branch] gh/malfet/502/head -> origin/gh/malfet/502/head 2025-09-07T07:28:18.8342561Z * [new branch] gh/malfet/502/orig -> origin/gh/malfet/502/orig 2025-09-07T07:28:18.8343809Z * [new branch] gh/malfet/503/base -> origin/gh/malfet/503/base 2025-09-07T07:28:18.8344635Z * [new branch] gh/malfet/503/head -> origin/gh/malfet/503/head 2025-09-07T07:28:18.8345475Z * [new branch] gh/malfet/503/orig -> origin/gh/malfet/503/orig 2025-09-07T07:28:18.8346691Z * [new branch] gh/malfet/504/base -> origin/gh/malfet/504/base 2025-09-07T07:28:18.8347612Z * [new branch] gh/malfet/504/head -> origin/gh/malfet/504/head 2025-09-07T07:28:18.8348449Z * [new branch] gh/malfet/504/orig -> origin/gh/malfet/504/orig 2025-09-07T07:28:18.8349688Z * [new branch] gh/malfet/505/base -> origin/gh/malfet/505/base 2025-09-07T07:28:18.8350509Z * [new branch] gh/malfet/505/head -> origin/gh/malfet/505/head 2025-09-07T07:28:18.8351390Z * [new branch] gh/malfet/505/orig -> origin/gh/malfet/505/orig 2025-09-07T07:28:18.8352724Z * [new branch] gh/malfet/506/base -> origin/gh/malfet/506/base 2025-09-07T07:28:18.8353479Z * [new branch] gh/malfet/506/head -> origin/gh/malfet/506/head 2025-09-07T07:28:18.8354308Z * [new branch] gh/malfet/506/orig -> origin/gh/malfet/506/orig 2025-09-07T07:28:18.8355494Z * [new branch] gh/malfet/507/base -> origin/gh/malfet/507/base 2025-09-07T07:28:18.8356371Z * [new branch] gh/malfet/507/head -> origin/gh/malfet/507/head 2025-09-07T07:28:18.8357177Z * [new branch] gh/malfet/507/orig -> origin/gh/malfet/507/orig 2025-09-07T07:28:18.8358505Z * [new branch] gh/malfet/508/base -> origin/gh/malfet/508/base 2025-09-07T07:28:18.8359406Z * [new branch] gh/malfet/508/head -> origin/gh/malfet/508/head 2025-09-07T07:28:18.8360317Z * [new branch] gh/malfet/508/orig -> origin/gh/malfet/508/orig 2025-09-07T07:28:18.8361710Z * [new branch] gh/malfet/509/base -> origin/gh/malfet/509/base 2025-09-07T07:28:18.8362579Z * [new branch] gh/malfet/509/head -> origin/gh/malfet/509/head 2025-09-07T07:28:18.8363496Z * [new branch] gh/malfet/509/orig -> origin/gh/malfet/509/orig 2025-09-07T07:28:18.8364848Z * [new branch] gh/malfet/510/base -> origin/gh/malfet/510/base 2025-09-07T07:28:18.8365801Z * [new branch] gh/malfet/510/head -> origin/gh/malfet/510/head 2025-09-07T07:28:18.8366709Z * [new branch] gh/malfet/510/orig -> origin/gh/malfet/510/orig 2025-09-07T07:28:18.8367882Z * [new branch] gh/malfet/511/base -> origin/gh/malfet/511/base 2025-09-07T07:28:18.8368743Z * [new branch] gh/malfet/511/head -> origin/gh/malfet/511/head 2025-09-07T07:28:18.8369591Z * [new branch] gh/malfet/511/orig -> origin/gh/malfet/511/orig 2025-09-07T07:28:18.8370744Z * [new branch] gh/malfet/512/base -> origin/gh/malfet/512/base 2025-09-07T07:28:18.8371603Z * [new branch] gh/malfet/512/head -> origin/gh/malfet/512/head 2025-09-07T07:28:18.8372470Z * [new branch] gh/malfet/512/orig -> origin/gh/malfet/512/orig 2025-09-07T07:28:18.8373694Z * [new branch] gh/malfet/513/base -> origin/gh/malfet/513/base 2025-09-07T07:28:18.8374567Z * [new branch] gh/malfet/513/head -> origin/gh/malfet/513/head 2025-09-07T07:28:18.8375386Z * [new branch] gh/malfet/513/orig -> origin/gh/malfet/513/orig 2025-09-07T07:28:18.8376579Z * [new branch] gh/malfet/64/base -> origin/gh/malfet/64/base 2025-09-07T07:28:18.8377432Z * [new branch] gh/malfet/64/head -> origin/gh/malfet/64/head 2025-09-07T07:28:18.8378936Z * [new branch] gh/manuelcandales/10/base -> origin/gh/manuelcandales/10/base 2025-09-07T07:28:18.8379766Z * [new branch] gh/manuelcandales/10/head -> origin/gh/manuelcandales/10/head 2025-09-07T07:28:18.8380606Z * [new branch] gh/manuelcandales/10/orig -> origin/gh/manuelcandales/10/orig 2025-09-07T07:28:18.8381732Z * [new branch] gh/manuelcandales/11/base -> origin/gh/manuelcandales/11/base 2025-09-07T07:28:18.8382537Z * [new branch] gh/manuelcandales/11/head -> origin/gh/manuelcandales/11/head 2025-09-07T07:28:18.8383343Z * [new branch] gh/manuelcandales/11/orig -> origin/gh/manuelcandales/11/orig 2025-09-07T07:28:18.8384589Z * [new branch] gh/manuelcandales/9/base -> origin/gh/manuelcandales/9/base 2025-09-07T07:28:18.8385391Z * [new branch] gh/manuelcandales/9/head -> origin/gh/manuelcandales/9/head 2025-09-07T07:28:18.8386336Z * [new branch] gh/manuelcandales/9/orig -> origin/gh/manuelcandales/9/orig 2025-09-07T07:28:18.8388039Z * [new branch] gh/markkm/1/base -> origin/gh/markkm/1/base 2025-09-07T07:28:18.8389900Z * [new branch] gh/masnesral/204/base -> origin/gh/masnesral/204/base 2025-09-07T07:28:18.8390849Z * [new branch] gh/masnesral/204/head -> origin/gh/masnesral/204/head 2025-09-07T07:28:18.8391723Z * [new branch] gh/masnesral/204/orig -> origin/gh/masnesral/204/orig 2025-09-07T07:28:18.8392993Z * [new branch] gh/masnesral/235/base -> origin/gh/masnesral/235/base 2025-09-07T07:28:18.8393942Z * [new branch] gh/masnesral/235/head -> origin/gh/masnesral/235/head 2025-09-07T07:28:18.8394832Z * [new branch] gh/masnesral/235/orig -> origin/gh/masnesral/235/orig 2025-09-07T07:28:18.8396079Z * [new branch] gh/masnesral/34/base -> origin/gh/masnesral/34/base 2025-09-07T07:28:18.8397688Z * [new branch] gh/mhorowitz/0/base -> origin/gh/mhorowitz/0/base 2025-09-07T07:28:18.8398599Z * [new branch] gh/mhorowitz/0/head -> origin/gh/mhorowitz/0/head 2025-09-07T07:28:18.8399606Z * [new branch] gh/mhorowitz/1/base -> origin/gh/mhorowitz/1/base 2025-09-07T07:28:18.8400480Z * [new branch] gh/mhorowitz/1/head -> origin/gh/mhorowitz/1/head 2025-09-07T07:28:18.8401523Z * [new branch] gh/mhorowitz/2/base -> origin/gh/mhorowitz/2/base 2025-09-07T07:28:18.8402404Z * [new branch] gh/mhorowitz/2/head -> origin/gh/mhorowitz/2/head 2025-09-07T07:28:18.8403470Z * [new branch] gh/mhorowitz/3/base -> origin/gh/mhorowitz/3/base 2025-09-07T07:28:18.8404253Z * [new branch] gh/mhorowitz/3/head -> origin/gh/mhorowitz/3/head 2025-09-07T07:28:18.8405290Z * [new branch] gh/mhorowitz/4/base -> origin/gh/mhorowitz/4/base 2025-09-07T07:28:18.8406095Z * [new branch] gh/mhorowitz/4/head -> origin/gh/mhorowitz/4/head 2025-09-07T07:28:18.8407191Z * [new branch] gh/mhorowitz/5/base -> origin/gh/mhorowitz/5/base 2025-09-07T07:28:18.8407948Z * [new branch] gh/mhorowitz/5/head -> origin/gh/mhorowitz/5/head 2025-09-07T07:28:18.8409019Z * [new branch] gh/mhorowitz/6/base -> origin/gh/mhorowitz/6/base 2025-09-07T07:28:18.8409839Z * [new branch] gh/mhorowitz/6/head -> origin/gh/mhorowitz/6/head 2025-09-07T07:28:18.8411348Z * [new branch] gh/mikaylagawarecki/234/base -> origin/gh/mikaylagawarecki/234/base 2025-09-07T07:28:18.8412407Z * [new branch] gh/mikaylagawarecki/234/head -> origin/gh/mikaylagawarecki/234/head 2025-09-07T07:28:18.8413570Z * [new branch] gh/mikaylagawarecki/235/base -> origin/gh/mikaylagawarecki/235/base 2025-09-07T07:28:18.8414711Z * [new branch] gh/mikaylagawarecki/235/head -> origin/gh/mikaylagawarecki/235/head 2025-09-07T07:28:18.8415493Z * [new branch] gh/mikaylagawarecki/236/base -> origin/gh/mikaylagawarecki/236/base 2025-09-07T07:28:18.8416279Z * [new branch] gh/mikaylagawarecki/236/head -> origin/gh/mikaylagawarecki/236/head 2025-09-07T07:28:18.8417385Z * [new branch] gh/mikaylagawarecki/237/base -> origin/gh/mikaylagawarecki/237/base 2025-09-07T07:28:18.8418159Z * [new branch] gh/mikaylagawarecki/237/head -> origin/gh/mikaylagawarecki/237/head 2025-09-07T07:28:18.8419405Z * [new branch] gh/mikaylagawarecki/238/base -> origin/gh/mikaylagawarecki/238/base 2025-09-07T07:28:18.8420357Z * [new branch] gh/mikaylagawarecki/238/head -> origin/gh/mikaylagawarecki/238/head 2025-09-07T07:28:18.8421482Z * [new branch] gh/mikaylagawarecki/317/base -> origin/gh/mikaylagawarecki/317/base 2025-09-07T07:28:18.8422376Z * [new branch] gh/mikaylagawarecki/317/head -> origin/gh/mikaylagawarecki/317/head 2025-09-07T07:28:18.8423364Z * [new branch] gh/mikaylagawarecki/317/orig -> origin/gh/mikaylagawarecki/317/orig 2025-09-07T07:28:18.8424482Z * [new branch] gh/mikaylagawarecki/320/base -> origin/gh/mikaylagawarecki/320/base 2025-09-07T07:28:18.8425302Z * [new branch] gh/mikaylagawarecki/320/head -> origin/gh/mikaylagawarecki/320/head 2025-09-07T07:28:18.8426167Z * [new branch] gh/mikaylagawarecki/320/orig -> origin/gh/mikaylagawarecki/320/orig 2025-09-07T07:28:18.8427445Z * [new branch] gh/mikaylagawarecki/329/base -> origin/gh/mikaylagawarecki/329/base 2025-09-07T07:28:18.8428282Z * [new branch] gh/mikaylagawarecki/329/head -> origin/gh/mikaylagawarecki/329/head 2025-09-07T07:28:18.8429156Z * [new branch] gh/mikaylagawarecki/329/orig -> origin/gh/mikaylagawarecki/329/orig 2025-09-07T07:28:18.8430460Z * [new branch] gh/mikaylagawarecki/330/base -> origin/gh/mikaylagawarecki/330/base 2025-09-07T07:28:18.8431346Z * [new branch] gh/mikaylagawarecki/330/head -> origin/gh/mikaylagawarecki/330/head 2025-09-07T07:28:18.8432159Z * [new branch] gh/mikaylagawarecki/330/orig -> origin/gh/mikaylagawarecki/330/orig 2025-09-07T07:28:18.8433365Z * [new branch] gh/mikaylagawarecki/331/base -> origin/gh/mikaylagawarecki/331/base 2025-09-07T07:28:18.8434281Z * [new branch] gh/mikaylagawarecki/331/head -> origin/gh/mikaylagawarecki/331/head 2025-09-07T07:28:18.8435106Z * [new branch] gh/mikaylagawarecki/331/orig -> origin/gh/mikaylagawarecki/331/orig 2025-09-07T07:28:18.8436478Z * [new branch] gh/mikaylagawarecki/332/base -> origin/gh/mikaylagawarecki/332/base 2025-09-07T07:28:18.8437319Z * [new branch] gh/mikaylagawarecki/332/head -> origin/gh/mikaylagawarecki/332/head 2025-09-07T07:28:18.8438096Z * [new branch] gh/mikaylagawarecki/332/orig -> origin/gh/mikaylagawarecki/332/orig 2025-09-07T07:28:18.8439341Z * [new branch] gh/mikaylagawarecki/334/base -> origin/gh/mikaylagawarecki/334/base 2025-09-07T07:28:18.8440181Z * [new branch] gh/mikaylagawarecki/334/head -> origin/gh/mikaylagawarecki/334/head 2025-09-07T07:28:18.8441056Z * [new branch] gh/mikaylagawarecki/334/orig -> origin/gh/mikaylagawarecki/334/orig 2025-09-07T07:28:18.8442282Z * [new branch] gh/mikaylagawarecki/335/base -> origin/gh/mikaylagawarecki/335/base 2025-09-07T07:28:18.8443187Z * [new branch] gh/mikaylagawarecki/335/head -> origin/gh/mikaylagawarecki/335/head 2025-09-07T07:28:18.8444000Z * [new branch] gh/mikaylagawarecki/335/orig -> origin/gh/mikaylagawarecki/335/orig 2025-09-07T07:28:18.8445180Z * [new branch] gh/mikaylagawarecki/336/base -> origin/gh/mikaylagawarecki/336/base 2025-09-07T07:28:18.8446073Z * [new branch] gh/mikaylagawarecki/336/head -> origin/gh/mikaylagawarecki/336/head 2025-09-07T07:28:18.8446931Z * [new branch] gh/mikaylagawarecki/336/orig -> origin/gh/mikaylagawarecki/336/orig 2025-09-07T07:28:18.8448042Z * [new branch] gh/mikaylagawarecki/337/base -> origin/gh/mikaylagawarecki/337/base 2025-09-07T07:28:18.8448881Z * [new branch] gh/mikaylagawarecki/337/head -> origin/gh/mikaylagawarecki/337/head 2025-09-07T07:28:18.8449736Z * [new branch] gh/mikaylagawarecki/337/orig -> origin/gh/mikaylagawarecki/337/orig 2025-09-07T07:28:18.8450886Z * [new branch] gh/mikaylagawarecki/338/base -> origin/gh/mikaylagawarecki/338/base 2025-09-07T07:28:18.8451809Z * [new branch] gh/mikaylagawarecki/338/head -> origin/gh/mikaylagawarecki/338/head 2025-09-07T07:28:18.8452652Z * [new branch] gh/mikaylagawarecki/338/orig -> origin/gh/mikaylagawarecki/338/orig 2025-09-07T07:28:18.8454376Z * [new branch] gh/mikaylagawarecki/339/base -> origin/gh/mikaylagawarecki/339/base 2025-09-07T07:28:18.8455264Z * [new branch] gh/mikaylagawarecki/339/head -> origin/gh/mikaylagawarecki/339/head 2025-09-07T07:28:18.8456182Z * [new branch] gh/mikaylagawarecki/339/orig -> origin/gh/mikaylagawarecki/339/orig 2025-09-07T07:28:18.8458002Z * [new branch] gh/mlazos/1/base -> origin/gh/mlazos/1/base 2025-09-07T07:28:18.8458941Z * [new branch] gh/mlazos/1/head -> origin/gh/mlazos/1/head 2025-09-07T07:28:18.8459749Z * [new branch] gh/mlazos/1/orig -> origin/gh/mlazos/1/orig 2025-09-07T07:28:18.8460956Z * [new branch] gh/mlazos/12/base -> origin/gh/mlazos/12/base 2025-09-07T07:28:18.8461805Z * [new branch] gh/mlazos/12/head -> origin/gh/mlazos/12/head 2025-09-07T07:28:18.8462664Z * [new branch] gh/mlazos/12/orig -> origin/gh/mlazos/12/orig 2025-09-07T07:28:18.8464108Z * [new branch] gh/mlazos/13/base -> origin/gh/mlazos/13/base 2025-09-07T07:28:18.8464986Z * [new branch] gh/mlazos/13/head -> origin/gh/mlazos/13/head 2025-09-07T07:28:18.8468487Z * [new branch] gh/mlazos/13/orig -> origin/gh/mlazos/13/orig 2025-09-07T07:28:18.8469821Z * [new branch] gh/mlazos/14/base -> origin/gh/mlazos/14/base 2025-09-07T07:28:18.8470663Z * [new branch] gh/mlazos/14/head -> origin/gh/mlazos/14/head 2025-09-07T07:28:18.8471554Z * [new branch] gh/mlazos/14/orig -> origin/gh/mlazos/14/orig 2025-09-07T07:28:18.8472869Z * [new branch] gh/mlazos/15/base -> origin/gh/mlazos/15/base 2025-09-07T07:28:18.8473713Z * [new branch] gh/mlazos/15/head -> origin/gh/mlazos/15/head 2025-09-07T07:28:18.8474523Z * [new branch] gh/mlazos/15/orig -> origin/gh/mlazos/15/orig 2025-09-07T07:28:18.8475820Z * [new branch] gh/mlazos/16/base -> origin/gh/mlazos/16/base 2025-09-07T07:28:18.8476724Z * [new branch] gh/mlazos/16/head -> origin/gh/mlazos/16/head 2025-09-07T07:28:18.8477588Z * [new branch] gh/mlazos/16/orig -> origin/gh/mlazos/16/orig 2025-09-07T07:28:18.8478674Z * [new branch] gh/mlazos/17/base -> origin/gh/mlazos/17/base 2025-09-07T07:28:18.8479579Z * [new branch] gh/mlazos/17/head -> origin/gh/mlazos/17/head 2025-09-07T07:28:18.8480369Z * [new branch] gh/mlazos/17/orig -> origin/gh/mlazos/17/orig 2025-09-07T07:28:18.8481714Z * [new branch] gh/mlazos/2/base -> origin/gh/mlazos/2/base 2025-09-07T07:28:18.8482468Z * [new branch] gh/mlazos/2/head -> origin/gh/mlazos/2/head 2025-09-07T07:28:18.8483327Z * [new branch] gh/mlazos/2/orig -> origin/gh/mlazos/2/orig 2025-09-07T07:28:18.8484594Z * [new branch] gh/mlazos/3/base -> origin/gh/mlazos/3/base 2025-09-07T07:28:18.8485404Z * [new branch] gh/mlazos/3/head -> origin/gh/mlazos/3/head 2025-09-07T07:28:18.8486229Z * [new branch] gh/mlazos/3/orig -> origin/gh/mlazos/3/orig 2025-09-07T07:28:18.8487836Z * [new branch] gh/mrmiywj/1/base -> origin/gh/mrmiywj/1/base 2025-09-07T07:28:18.8488763Z * [new branch] gh/mrmiywj/1/head -> origin/gh/mrmiywj/1/head 2025-09-07T07:28:18.8490240Z * [new branch] gh/muchulee8/62/base -> origin/gh/muchulee8/62/base 2025-09-07T07:28:18.8491612Z * [new branch] gh/muchulee8/62/head -> origin/gh/muchulee8/62/head 2025-09-07T07:28:18.8492573Z * [new branch] gh/muchulee8/62/orig -> origin/gh/muchulee8/62/orig 2025-09-07T07:28:18.8493772Z * [new branch] gh/muchulee8/63/base -> origin/gh/muchulee8/63/base 2025-09-07T07:28:18.8494669Z * [new branch] gh/muchulee8/63/head -> origin/gh/muchulee8/63/head 2025-09-07T07:28:18.8495674Z * [new branch] gh/muchulee8/63/orig -> origin/gh/muchulee8/63/orig 2025-09-07T07:28:18.8496929Z * [new branch] gh/muchulee8/64/base -> origin/gh/muchulee8/64/base 2025-09-07T07:28:18.8497752Z * [new branch] gh/muchulee8/64/head -> origin/gh/muchulee8/64/head 2025-09-07T07:28:18.8498600Z * [new branch] gh/muchulee8/64/orig -> origin/gh/muchulee8/64/orig 2025-09-07T07:28:18.8499871Z * [new branch] gh/muchulee8/65/base -> origin/gh/muchulee8/65/base 2025-09-07T07:28:18.8500810Z * [new branch] gh/muchulee8/65/head -> origin/gh/muchulee8/65/head 2025-09-07T07:28:18.8501822Z * [new branch] gh/muchulee8/65/orig -> origin/gh/muchulee8/65/orig 2025-09-07T07:28:18.8503552Z * [new branch] gh/naveenthangudu/1/base -> origin/gh/naveenthangudu/1/base 2025-09-07T07:28:18.8504503Z * [new branch] gh/naveenthangudu/1/head -> origin/gh/naveenthangudu/1/head 2025-09-07T07:28:18.8505462Z * [new branch] gh/naveenthangudu/1/orig -> origin/gh/naveenthangudu/1/orig 2025-09-07T07:28:18.8506746Z * [new branch] gh/naveenthangudu/2/base -> origin/gh/naveenthangudu/2/base 2025-09-07T07:28:18.8507689Z * [new branch] gh/naveenthangudu/2/head -> origin/gh/naveenthangudu/2/head 2025-09-07T07:28:18.8508631Z * [new branch] gh/naveenthangudu/2/orig -> origin/gh/naveenthangudu/2/orig 2025-09-07T07:28:18.8509806Z * [new branch] gh/naveenthangudu/3/base -> origin/gh/naveenthangudu/3/base 2025-09-07T07:28:18.8510717Z * [new branch] gh/naveenthangudu/3/head -> origin/gh/naveenthangudu/3/head 2025-09-07T07:28:18.8511610Z * [new branch] gh/naveenthangudu/3/orig -> origin/gh/naveenthangudu/3/orig 2025-09-07T07:28:18.8512687Z * [new branch] gh/naveenthangudu/4/base -> origin/gh/naveenthangudu/4/base 2025-09-07T07:28:18.8513551Z * [new branch] gh/naveenthangudu/4/head -> origin/gh/naveenthangudu/4/head 2025-09-07T07:28:18.8514508Z * [new branch] gh/naveenthangudu/4/orig -> origin/gh/naveenthangudu/4/orig 2025-09-07T07:28:18.8515655Z * [new branch] gh/naveenthangudu/5/base -> origin/gh/naveenthangudu/5/base 2025-09-07T07:28:18.8516585Z * [new branch] gh/naveenthangudu/5/head -> origin/gh/naveenthangudu/5/head 2025-09-07T07:28:18.8517682Z * [new branch] gh/naveenthangudu/5/orig -> origin/gh/naveenthangudu/5/orig 2025-09-07T07:28:18.8518924Z * [new branch] gh/naveenthangudu/6/base -> origin/gh/naveenthangudu/6/base 2025-09-07T07:28:18.8519780Z * [new branch] gh/naveenthangudu/6/head -> origin/gh/naveenthangudu/6/head 2025-09-07T07:28:18.8520563Z * [new branch] gh/naveenthangudu/6/orig -> origin/gh/naveenthangudu/6/orig 2025-09-07T07:28:18.8522029Z * [new branch] gh/oulgen/35/base -> origin/gh/oulgen/35/base 2025-09-07T07:28:18.8522871Z * [new branch] gh/oulgen/35/head -> origin/gh/oulgen/35/head 2025-09-07T07:28:18.8523736Z * [new branch] gh/oulgen/35/orig -> origin/gh/oulgen/35/orig 2025-09-07T07:28:18.8524934Z * [new branch] gh/oulgen/48/base -> origin/gh/oulgen/48/base 2025-09-07T07:28:18.8525790Z * [new branch] gh/oulgen/48/head -> origin/gh/oulgen/48/head 2025-09-07T07:28:18.8526618Z * [new branch] gh/oulgen/48/orig -> origin/gh/oulgen/48/orig 2025-09-07T07:28:18.8527737Z * [new branch] gh/oulgen/49/base -> origin/gh/oulgen/49/base 2025-09-07T07:28:18.8528692Z * [new branch] gh/oulgen/49/head -> origin/gh/oulgen/49/head 2025-09-07T07:28:18.8529578Z * [new branch] gh/oulgen/49/orig -> origin/gh/oulgen/49/orig 2025-09-07T07:28:18.8531155Z * [new branch] gh/pearu/108/base -> origin/gh/pearu/108/base 2025-09-07T07:28:18.8532195Z * [new branch] gh/pearu/108/head -> origin/gh/pearu/108/head 2025-09-07T07:28:18.8533101Z * [new branch] gh/pearu/108/orig -> origin/gh/pearu/108/orig 2025-09-07T07:28:18.8534516Z * [new branch] gh/pearu/109/base -> origin/gh/pearu/109/base 2025-09-07T07:28:18.8535363Z * [new branch] gh/pearu/109/head -> origin/gh/pearu/109/head 2025-09-07T07:28:18.8536245Z * [new branch] gh/pearu/109/orig -> origin/gh/pearu/109/orig 2025-09-07T07:28:18.8537410Z * [new branch] gh/pearu/110/base -> origin/gh/pearu/110/base 2025-09-07T07:28:18.8538361Z * [new branch] gh/pearu/110/head -> origin/gh/pearu/110/head 2025-09-07T07:28:18.8539206Z * [new branch] gh/pearu/110/orig -> origin/gh/pearu/110/orig 2025-09-07T07:28:18.8540512Z * [new branch] gh/pearu/111/base -> origin/gh/pearu/111/base 2025-09-07T07:28:18.8541368Z * [new branch] gh/pearu/111/head -> origin/gh/pearu/111/head 2025-09-07T07:28:18.8542215Z * [new branch] gh/pearu/111/orig -> origin/gh/pearu/111/orig 2025-09-07T07:28:18.8543325Z * [new branch] gh/pearu/112/base -> origin/gh/pearu/112/base 2025-09-07T07:28:18.8544206Z * [new branch] gh/pearu/112/head -> origin/gh/pearu/112/head 2025-09-07T07:28:18.8545064Z * [new branch] gh/pearu/112/orig -> origin/gh/pearu/112/orig 2025-09-07T07:28:18.8546268Z * [new branch] gh/pearu/113/base -> origin/gh/pearu/113/base 2025-09-07T07:28:18.8547111Z * [new branch] gh/pearu/113/head -> origin/gh/pearu/113/head 2025-09-07T07:28:18.8547928Z * [new branch] gh/pearu/113/orig -> origin/gh/pearu/113/orig 2025-09-07T07:28:18.8549156Z * [new branch] gh/pearu/114/base -> origin/gh/pearu/114/base 2025-09-07T07:28:18.8550000Z * [new branch] gh/pearu/114/head -> origin/gh/pearu/114/head 2025-09-07T07:28:18.8550964Z * [new branch] gh/pearu/114/orig -> origin/gh/pearu/114/orig 2025-09-07T07:28:18.8552284Z * [new branch] gh/pearu/115/base -> origin/gh/pearu/115/base 2025-09-07T07:28:18.8553134Z * [new branch] gh/pearu/115/head -> origin/gh/pearu/115/head 2025-09-07T07:28:18.8553908Z * [new branch] gh/pearu/115/orig -> origin/gh/pearu/115/orig 2025-09-07T07:28:18.8555112Z * [new branch] gh/pearu/116/base -> origin/gh/pearu/116/base 2025-09-07T07:28:18.8555912Z * [new branch] gh/pearu/116/head -> origin/gh/pearu/116/head 2025-09-07T07:28:18.8557103Z * [new branch] gh/pearu/116/orig -> origin/gh/pearu/116/orig 2025-09-07T07:28:18.8558283Z * [new branch] gh/pearu/117/base -> origin/gh/pearu/117/base 2025-09-07T07:28:18.8559102Z * [new branch] gh/pearu/117/head -> origin/gh/pearu/117/head 2025-09-07T07:28:18.8559823Z * [new branch] gh/pearu/117/orig -> origin/gh/pearu/117/orig 2025-09-07T07:28:18.8561433Z * [new branch] gh/pearu/56/base -> origin/gh/pearu/56/base 2025-09-07T07:28:18.8562457Z * [new branch] gh/pearu/56/head -> origin/gh/pearu/56/head 2025-09-07T07:28:18.8563334Z * [new branch] gh/pearu/56/orig -> origin/gh/pearu/56/orig 2025-09-07T07:28:18.8565092Z * [new branch] gh/pearu/97/base -> origin/gh/pearu/97/base 2025-09-07T07:28:18.8566045Z * [new branch] gh/pearu/97/head -> origin/gh/pearu/97/head 2025-09-07T07:28:18.8566885Z * [new branch] gh/pearu/97/orig -> origin/gh/pearu/97/orig 2025-09-07T07:28:18.8568410Z * [new branch] gh/qqaatw/29/base -> origin/gh/qqaatw/29/base 2025-09-07T07:28:18.8569493Z * [new branch] gh/qqaatw/29/head -> origin/gh/qqaatw/29/head 2025-09-07T07:28:18.8570153Z * [new branch] gh/qqaatw/29/orig -> origin/gh/qqaatw/29/orig 2025-09-07T07:28:18.8571373Z * [new branch] gh/raymo/refresh-script -> origin/gh/raymo/refresh-script 2025-09-07T07:28:18.8572702Z * [new branch] gh/rec/141/base -> origin/gh/rec/141/base 2025-09-07T07:28:18.8573635Z * [new branch] gh/rec/141/head -> origin/gh/rec/141/head 2025-09-07T07:28:18.8574817Z * [new branch] gh/rec/153/base -> origin/gh/rec/153/base 2025-09-07T07:28:18.8575655Z * [new branch] gh/rec/153/head -> origin/gh/rec/153/head 2025-09-07T07:28:18.8576511Z * [new branch] gh/rec/153/orig -> origin/gh/rec/153/orig 2025-09-07T07:28:18.8577650Z * [new branch] gh/rec/154/base -> origin/gh/rec/154/base 2025-09-07T07:28:18.8578517Z * [new branch] gh/rec/154/head -> origin/gh/rec/154/head 2025-09-07T07:28:18.8579371Z * [new branch] gh/rec/154/orig -> origin/gh/rec/154/orig 2025-09-07T07:28:18.8580548Z * [new branch] gh/rec/156/base -> origin/gh/rec/156/base 2025-09-07T07:28:18.8581399Z * [new branch] gh/rec/156/head -> origin/gh/rec/156/head 2025-09-07T07:28:18.8582260Z * [new branch] gh/rec/156/orig -> origin/gh/rec/156/orig 2025-09-07T07:28:18.8583407Z * [new branch] gh/rec/160/base -> origin/gh/rec/160/base 2025-09-07T07:28:18.8584273Z * [new branch] gh/rec/160/head -> origin/gh/rec/160/head 2025-09-07T07:28:18.8585119Z * [new branch] gh/rec/160/orig -> origin/gh/rec/160/orig 2025-09-07T07:28:18.8586352Z * [new branch] gh/rec/162/base -> origin/gh/rec/162/base 2025-09-07T07:28:18.8587208Z * [new branch] gh/rec/162/head -> origin/gh/rec/162/head 2025-09-07T07:28:18.8588027Z * [new branch] gh/rec/162/orig -> origin/gh/rec/162/orig 2025-09-07T07:28:18.8589229Z * [new branch] gh/rec/163/base -> origin/gh/rec/163/base 2025-09-07T07:28:18.8590057Z * [new branch] gh/rec/163/head -> origin/gh/rec/163/head 2025-09-07T07:28:18.8590886Z * [new branch] gh/rec/163/orig -> origin/gh/rec/163/orig 2025-09-07T07:28:18.8592015Z * [new branch] gh/rec/164/base -> origin/gh/rec/164/base 2025-09-07T07:28:18.8592860Z * [new branch] gh/rec/164/head -> origin/gh/rec/164/head 2025-09-07T07:28:18.8593731Z * [new branch] gh/rec/164/orig -> origin/gh/rec/164/orig 2025-09-07T07:28:18.8594920Z * [new branch] gh/rec/165/base -> origin/gh/rec/165/base 2025-09-07T07:28:18.8595738Z * [new branch] gh/rec/165/head -> origin/gh/rec/165/head 2025-09-07T07:28:18.8596617Z * [new branch] gh/rec/165/orig -> origin/gh/rec/165/orig 2025-09-07T07:28:18.8597780Z * [new branch] gh/rec/166/base -> origin/gh/rec/166/base 2025-09-07T07:28:18.8598668Z * [new branch] gh/rec/166/head -> origin/gh/rec/166/head 2025-09-07T07:28:18.8599462Z * [new branch] gh/rec/166/orig -> origin/gh/rec/166/orig 2025-09-07T07:28:18.8601017Z * [new branch] gh/robert-hardwick/1/base -> origin/gh/robert-hardwick/1/base 2025-09-07T07:28:18.8601904Z * [new branch] gh/robert-hardwick/1/head -> origin/gh/robert-hardwick/1/head 2025-09-07T07:28:18.8602757Z * [new branch] gh/robert-hardwick/1/orig -> origin/gh/robert-hardwick/1/orig 2025-09-07T07:28:18.8603952Z * [new branch] gh/robert-hardwick/2/base -> origin/gh/robert-hardwick/2/base 2025-09-07T07:28:18.8604967Z * [new branch] gh/robert-hardwick/2/head -> origin/gh/robert-hardwick/2/head 2025-09-07T07:28:18.8605719Z * [new branch] gh/robert-hardwick/2/orig -> origin/gh/robert-hardwick/2/orig 2025-09-07T07:28:18.8606982Z * [new branch] gh/robert-hardwick/3/base -> origin/gh/robert-hardwick/3/base 2025-09-07T07:28:18.8607839Z * [new branch] gh/robert-hardwick/3/head -> origin/gh/robert-hardwick/3/head 2025-09-07T07:28:18.8608679Z * [new branch] gh/robert-hardwick/3/orig -> origin/gh/robert-hardwick/3/orig 2025-09-07T07:28:18.8609876Z * [new branch] gh/robert-hardwick/4/base -> origin/gh/robert-hardwick/4/base 2025-09-07T07:28:18.8610709Z * [new branch] gh/robert-hardwick/4/head -> origin/gh/robert-hardwick/4/head 2025-09-07T07:28:18.8611558Z * [new branch] gh/robert-hardwick/4/orig -> origin/gh/robert-hardwick/4/orig 2025-09-07T07:28:18.8612967Z * [new branch] gh/rtimpe/1/base -> origin/gh/rtimpe/1/base 2025-09-07T07:28:18.8613876Z * [new branch] gh/rtimpe/1/head -> origin/gh/rtimpe/1/head 2025-09-07T07:28:18.8615027Z * [new branch] gh/rtimpe/10/base -> origin/gh/rtimpe/10/base 2025-09-07T07:28:18.8615864Z * [new branch] gh/rtimpe/10/head -> origin/gh/rtimpe/10/head 2025-09-07T07:28:18.8616767Z * [new branch] gh/rtimpe/10/orig -> origin/gh/rtimpe/10/orig 2025-09-07T07:28:18.8617918Z * [new branch] gh/rtimpe/11/base -> origin/gh/rtimpe/11/base 2025-09-07T07:28:18.8618823Z * [new branch] gh/rtimpe/11/head -> origin/gh/rtimpe/11/head 2025-09-07T07:28:18.8619664Z * [new branch] gh/rtimpe/11/orig -> origin/gh/rtimpe/11/orig 2025-09-07T07:28:18.8620864Z * [new branch] gh/rtimpe/12/base -> origin/gh/rtimpe/12/base 2025-09-07T07:28:18.8621677Z * [new branch] gh/rtimpe/12/head -> origin/gh/rtimpe/12/head 2025-09-07T07:28:18.8622626Z * [new branch] gh/rtimpe/12/orig -> origin/gh/rtimpe/12/orig 2025-09-07T07:28:18.8623730Z * [new branch] gh/rtimpe/13/base -> origin/gh/rtimpe/13/base 2025-09-07T07:28:18.8624595Z * [new branch] gh/rtimpe/13/head -> origin/gh/rtimpe/13/head 2025-09-07T07:28:18.8625440Z * [new branch] gh/rtimpe/13/orig -> origin/gh/rtimpe/13/orig 2025-09-07T07:28:18.8626645Z * [new branch] gh/rtimpe/14/base -> origin/gh/rtimpe/14/base 2025-09-07T07:28:18.8627482Z * [new branch] gh/rtimpe/14/head -> origin/gh/rtimpe/14/head 2025-09-07T07:28:18.8628320Z * [new branch] gh/rtimpe/14/orig -> origin/gh/rtimpe/14/orig 2025-09-07T07:28:18.8629489Z * [new branch] gh/rtimpe/15/base -> origin/gh/rtimpe/15/base 2025-09-07T07:28:18.8630343Z * [new branch] gh/rtimpe/15/head -> origin/gh/rtimpe/15/head 2025-09-07T07:28:18.8631220Z * [new branch] gh/rtimpe/15/orig -> origin/gh/rtimpe/15/orig 2025-09-07T07:28:18.8632345Z * [new branch] gh/rtimpe/2/base -> origin/gh/rtimpe/2/base 2025-09-07T07:28:18.8633131Z * [new branch] gh/rtimpe/2/head -> origin/gh/rtimpe/2/head 2025-09-07T07:28:18.8634253Z * [new branch] gh/rtimpe/3/base -> origin/gh/rtimpe/3/base 2025-09-07T07:28:18.8635008Z * [new branch] gh/rtimpe/3/head -> origin/gh/rtimpe/3/head 2025-09-07T07:28:18.8636125Z * [new branch] gh/rtimpe/4/base -> origin/gh/rtimpe/4/base 2025-09-07T07:28:18.8636987Z * [new branch] gh/rtimpe/4/head -> origin/gh/rtimpe/4/head 2025-09-07T07:28:18.8638188Z * [new branch] gh/rtimpe/9/base -> origin/gh/rtimpe/9/base 2025-09-07T07:28:18.8638981Z * [new branch] gh/rtimpe/9/head -> origin/gh/rtimpe/9/head 2025-09-07T07:28:18.8639952Z * [new branch] gh/rtimpe/9/orig -> origin/gh/rtimpe/9/orig 2025-09-07T07:28:18.8641635Z * [new branch] gh/ruisizhang123/1/base -> origin/gh/ruisizhang123/1/base 2025-09-07T07:28:18.8642546Z * [new branch] gh/ruisizhang123/1/head -> origin/gh/ruisizhang123/1/head 2025-09-07T07:28:18.8643397Z * [new branch] gh/ruisizhang123/1/orig -> origin/gh/ruisizhang123/1/orig 2025-09-07T07:28:18.8644567Z * [new branch] gh/ruisizhang123/4/base -> origin/gh/ruisizhang123/4/base 2025-09-07T07:28:18.8645455Z * [new branch] gh/ruisizhang123/4/head -> origin/gh/ruisizhang123/4/head 2025-09-07T07:28:18.8646291Z * [new branch] gh/ruisizhang123/4/orig -> origin/gh/ruisizhang123/4/orig 2025-09-07T07:28:18.8647475Z * [new branch] gh/ruisizhang123/5/base -> origin/gh/ruisizhang123/5/base 2025-09-07T07:28:18.8648286Z * [new branch] gh/ruisizhang123/5/head -> origin/gh/ruisizhang123/5/head 2025-09-07T07:28:18.8649158Z * [new branch] gh/ruisizhang123/5/orig -> origin/gh/ruisizhang123/5/orig 2025-09-07T07:28:18.8650285Z * [new branch] gh/ruisizhang123/6/base -> origin/gh/ruisizhang123/6/base 2025-09-07T07:28:18.8651125Z * [new branch] gh/ruisizhang123/6/head -> origin/gh/ruisizhang123/6/head 2025-09-07T07:28:18.8652047Z * [new branch] gh/ruisizhang123/6/orig -> origin/gh/ruisizhang123/6/orig 2025-09-07T07:28:18.8653226Z * [new branch] gh/ruisizhang123/7/base -> origin/gh/ruisizhang123/7/base 2025-09-07T07:28:18.8654189Z * [new branch] gh/ruisizhang123/7/head -> origin/gh/ruisizhang123/7/head 2025-09-07T07:28:18.8655022Z * [new branch] gh/ruisizhang123/7/orig -> origin/gh/ruisizhang123/7/orig 2025-09-07T07:28:18.8656105Z * [new branch] gh/ruisizhang123/8/base -> origin/gh/ruisizhang123/8/base 2025-09-07T07:28:18.8657003Z * [new branch] gh/ruisizhang123/8/head -> origin/gh/ruisizhang123/8/head 2025-09-07T07:28:18.8657831Z * [new branch] gh/ruisizhang123/8/orig -> origin/gh/ruisizhang123/8/orig 2025-09-07T07:28:18.8658989Z * [new branch] gh/ruisizhang123/9/base -> origin/gh/ruisizhang123/9/base 2025-09-07T07:28:18.8659843Z * [new branch] gh/ruisizhang123/9/head -> origin/gh/ruisizhang123/9/head 2025-09-07T07:28:18.8660727Z * [new branch] gh/ruisizhang123/9/orig -> origin/gh/ruisizhang123/9/orig 2025-09-07T07:28:18.8662134Z * [new branch] gh/sarckk/2/base -> origin/gh/sarckk/2/base 2025-09-07T07:28:18.8662999Z * [new branch] gh/sarckk/2/head -> origin/gh/sarckk/2/head 2025-09-07T07:28:18.8663830Z * [new branch] gh/sarckk/2/orig -> origin/gh/sarckk/2/orig 2025-09-07T07:28:18.8665377Z * [new branch] gh/seemethere/35/base -> origin/gh/seemethere/35/base 2025-09-07T07:28:18.8666345Z * [new branch] gh/seemethere/35/head -> origin/gh/seemethere/35/head 2025-09-07T07:28:18.8667243Z * [new branch] gh/seemethere/35/orig -> origin/gh/seemethere/35/orig 2025-09-07T07:28:18.8668415Z * [new branch] gh/seemethere/37/base -> origin/gh/seemethere/37/base 2025-09-07T07:28:18.8669258Z * [new branch] gh/seemethere/37/head -> origin/gh/seemethere/37/head 2025-09-07T07:28:18.8670064Z * [new branch] gh/seemethere/37/orig -> origin/gh/seemethere/37/orig 2025-09-07T07:28:18.8671291Z * [new branch] gh/seemethere/43/base -> origin/gh/seemethere/43/base 2025-09-07T07:28:18.8672119Z * [new branch] gh/seemethere/43/head -> origin/gh/seemethere/43/head 2025-09-07T07:28:18.8673009Z * [new branch] gh/seemethere/43/orig -> origin/gh/seemethere/43/orig 2025-09-07T07:28:18.8674451Z * [new branch] gh/seemethere/44/base -> origin/gh/seemethere/44/base 2025-09-07T07:28:18.8675161Z * [new branch] gh/seemethere/44/head -> origin/gh/seemethere/44/head 2025-09-07T07:28:18.8676022Z * [new branch] gh/seemethere/44/orig -> origin/gh/seemethere/44/orig 2025-09-07T07:28:18.8677478Z * [new branch] gh/seemethere/48/base -> origin/gh/seemethere/48/base 2025-09-07T07:28:18.8678377Z * [new branch] gh/seemethere/48/head -> origin/gh/seemethere/48/head 2025-09-07T07:28:18.8679243Z * [new branch] gh/seemethere/48/orig -> origin/gh/seemethere/48/orig 2025-09-07T07:28:18.8680643Z * [new branch] gh/seemethere/49/base -> origin/gh/seemethere/49/base 2025-09-07T07:28:18.8681534Z * [new branch] gh/seemethere/49/head -> origin/gh/seemethere/49/head 2025-09-07T07:28:18.8682390Z * [new branch] gh/seemethere/49/orig -> origin/gh/seemethere/49/orig 2025-09-07T07:28:18.8683601Z * [new branch] gh/seemethere/52/base -> origin/gh/seemethere/52/base 2025-09-07T07:28:18.8684496Z * [new branch] gh/seemethere/52/head -> origin/gh/seemethere/52/head 2025-09-07T07:28:18.8685391Z * [new branch] gh/seemethere/52/orig -> origin/gh/seemethere/52/orig 2025-09-07T07:28:18.8686551Z * [new branch] gh/seemethere/53/base -> origin/gh/seemethere/53/base 2025-09-07T07:28:18.8687334Z * [new branch] gh/seemethere/53/head -> origin/gh/seemethere/53/head 2025-09-07T07:28:18.8688206Z * [new branch] gh/seemethere/53/orig -> origin/gh/seemethere/53/orig 2025-09-07T07:28:18.8689407Z * [new branch] gh/seemethere/54/base -> origin/gh/seemethere/54/base 2025-09-07T07:28:18.8690292Z * [new branch] gh/seemethere/54/head -> origin/gh/seemethere/54/head 2025-09-07T07:28:18.8691218Z * [new branch] gh/seemethere/54/orig -> origin/gh/seemethere/54/orig 2025-09-07T07:28:18.8692291Z * [new branch] gh/seemethere/55/base -> origin/gh/seemethere/55/base 2025-09-07T07:28:18.8693049Z * [new branch] gh/seemethere/55/head -> origin/gh/seemethere/55/head 2025-09-07T07:28:18.8693915Z * [new branch] gh/seemethere/55/orig -> origin/gh/seemethere/55/orig 2025-09-07T07:28:18.8695087Z * [new branch] gh/seemethere/56/base -> origin/gh/seemethere/56/base 2025-09-07T07:28:18.8695949Z * [new branch] gh/seemethere/56/head -> origin/gh/seemethere/56/head 2025-09-07T07:28:18.8696776Z * [new branch] gh/seemethere/56/orig -> origin/gh/seemethere/56/orig 2025-09-07T07:28:18.8697956Z * [new branch] gh/seemethere/57/base -> origin/gh/seemethere/57/base 2025-09-07T07:28:18.8698805Z * [new branch] gh/seemethere/57/head -> origin/gh/seemethere/57/head 2025-09-07T07:28:18.8699686Z * [new branch] gh/seemethere/57/orig -> origin/gh/seemethere/57/orig 2025-09-07T07:28:18.8700848Z * [new branch] gh/seemethere/58/base -> origin/gh/seemethere/58/base 2025-09-07T07:28:18.8701733Z * [new branch] gh/seemethere/58/head -> origin/gh/seemethere/58/head 2025-09-07T07:28:18.8702540Z * [new branch] gh/seemethere/58/orig -> origin/gh/seemethere/58/orig 2025-09-07T07:28:18.8703635Z * [new branch] gh/seemethere/59/base -> origin/gh/seemethere/59/base 2025-09-07T07:28:18.8704520Z * [new branch] gh/seemethere/59/head -> origin/gh/seemethere/59/head 2025-09-07T07:28:18.8705360Z * [new branch] gh/seemethere/59/orig -> origin/gh/seemethere/59/orig 2025-09-07T07:28:18.8706518Z * [new branch] gh/seemethere/60/base -> origin/gh/seemethere/60/base 2025-09-07T07:28:18.8707374Z * [new branch] gh/seemethere/60/head -> origin/gh/seemethere/60/head 2025-09-07T07:28:18.8708383Z * [new branch] gh/seemethere/60/orig -> origin/gh/seemethere/60/orig 2025-09-07T07:28:18.8709448Z * [new branch] gh/seemethere/61/base -> origin/gh/seemethere/61/base 2025-09-07T07:28:18.8710251Z * [new branch] gh/seemethere/61/head -> origin/gh/seemethere/61/head 2025-09-07T07:28:18.8711121Z * [new branch] gh/seemethere/61/orig -> origin/gh/seemethere/61/orig 2025-09-07T07:28:18.8712341Z * [new branch] gh/seemethere/62/base -> origin/gh/seemethere/62/base 2025-09-07T07:28:18.8713195Z * [new branch] gh/seemethere/62/head -> origin/gh/seemethere/62/head 2025-09-07T07:28:18.8714086Z * [new branch] gh/seemethere/62/orig -> origin/gh/seemethere/62/orig 2025-09-07T07:28:18.8715196Z * [new branch] gh/seemethere/63/base -> origin/gh/seemethere/63/base 2025-09-07T07:28:18.8716066Z * [new branch] gh/seemethere/63/head -> origin/gh/seemethere/63/head 2025-09-07T07:28:18.8716940Z * [new branch] gh/seemethere/63/orig -> origin/gh/seemethere/63/orig 2025-09-07T07:28:18.8718651Z * [new branch] gh/shunting314/145/base -> origin/gh/shunting314/145/base 2025-09-07T07:28:18.8719592Z * [new branch] gh/shunting314/145/head -> origin/gh/shunting314/145/head 2025-09-07T07:28:18.8720534Z * [new branch] gh/shunting314/145/orig -> origin/gh/shunting314/145/orig 2025-09-07T07:28:18.8721918Z * [new branch] gh/shunting314/176/base -> origin/gh/shunting314/176/base 2025-09-07T07:28:18.8722962Z * [new branch] gh/shunting314/176/head -> origin/gh/shunting314/176/head 2025-09-07T07:28:18.8724161Z * [new branch] gh/shunting314/176/orig -> origin/gh/shunting314/176/orig 2025-09-07T07:28:18.8725592Z * [new branch] gh/shunting314/211/base -> origin/gh/shunting314/211/base 2025-09-07T07:28:18.8726491Z * [new branch] gh/shunting314/211/head -> origin/gh/shunting314/211/head 2025-09-07T07:28:18.8727341Z * [new branch] gh/shunting314/211/orig -> origin/gh/shunting314/211/orig 2025-09-07T07:28:18.8728428Z * [new branch] gh/shunting314/212/base -> origin/gh/shunting314/212/base 2025-09-07T07:28:18.8729291Z * [new branch] gh/shunting314/212/head -> origin/gh/shunting314/212/head 2025-09-07T07:28:18.8730118Z * [new branch] gh/shunting314/212/orig -> origin/gh/shunting314/212/orig 2025-09-07T07:28:18.8731283Z * [new branch] gh/shunting314/213/base -> origin/gh/shunting314/213/base 2025-09-07T07:28:18.8732246Z * [new branch] gh/shunting314/213/head -> origin/gh/shunting314/213/head 2025-09-07T07:28:18.8733080Z * [new branch] gh/shunting314/213/orig -> origin/gh/shunting314/213/orig 2025-09-07T07:28:18.8734256Z * [new branch] gh/shunting314/214/base -> origin/gh/shunting314/214/base 2025-09-07T07:28:18.8735104Z * [new branch] gh/shunting314/214/head -> origin/gh/shunting314/214/head 2025-09-07T07:28:18.8735962Z * [new branch] gh/shunting314/214/orig -> origin/gh/shunting314/214/orig 2025-09-07T07:28:18.8737395Z * [new branch] gh/shunting314/215/base -> origin/gh/shunting314/215/base 2025-09-07T07:28:18.8738227Z * [new branch] gh/shunting314/215/head -> origin/gh/shunting314/215/head 2025-09-07T07:28:18.8739073Z * [new branch] gh/shunting314/215/orig -> origin/gh/shunting314/215/orig 2025-09-07T07:28:18.8740234Z * [new branch] gh/shunting314/216/base -> origin/gh/shunting314/216/base 2025-09-07T07:28:18.8741099Z * [new branch] gh/shunting314/216/head -> origin/gh/shunting314/216/head 2025-09-07T07:28:18.8741946Z * [new branch] gh/shunting314/216/orig -> origin/gh/shunting314/216/orig 2025-09-07T07:28:18.8743203Z * [new branch] gh/shunting314/217/base -> origin/gh/shunting314/217/base 2025-09-07T07:28:18.8744003Z * [new branch] gh/shunting314/217/head -> origin/gh/shunting314/217/head 2025-09-07T07:28:18.8744835Z * [new branch] gh/shunting314/217/orig -> origin/gh/shunting314/217/orig 2025-09-07T07:28:18.8746121Z * [new branch] gh/shunting314/218/base -> origin/gh/shunting314/218/base 2025-09-07T07:28:18.8746926Z * [new branch] gh/shunting314/218/head -> origin/gh/shunting314/218/head 2025-09-07T07:28:18.8747877Z * [new branch] gh/shunting314/218/orig -> origin/gh/shunting314/218/orig 2025-09-07T07:28:18.8748920Z * [new branch] gh/shunting314/219/base -> origin/gh/shunting314/219/base 2025-09-07T07:28:18.8749786Z * [new branch] gh/shunting314/219/head -> origin/gh/shunting314/219/head 2025-09-07T07:28:18.8750581Z * [new branch] gh/shunting314/219/orig -> origin/gh/shunting314/219/orig 2025-09-07T07:28:18.8751895Z * [new branch] gh/shunting314/220/base -> origin/gh/shunting314/220/base 2025-09-07T07:28:18.8752939Z * [new branch] gh/shunting314/220/head -> origin/gh/shunting314/220/head 2025-09-07T07:28:18.8753782Z * [new branch] gh/shunting314/220/orig -> origin/gh/shunting314/220/orig 2025-09-07T07:28:18.8754892Z * [new branch] gh/shunting314/221/base -> origin/gh/shunting314/221/base 2025-09-07T07:28:18.8755797Z * [new branch] gh/shunting314/221/head -> origin/gh/shunting314/221/head 2025-09-07T07:28:18.8756637Z * [new branch] gh/shunting314/221/orig -> origin/gh/shunting314/221/orig 2025-09-07T07:28:18.8757753Z * [new branch] gh/shunting314/222/base -> origin/gh/shunting314/222/base 2025-09-07T07:28:18.8758576Z * [new branch] gh/shunting314/222/head -> origin/gh/shunting314/222/head 2025-09-07T07:28:18.8759488Z * [new branch] gh/shunting314/222/orig -> origin/gh/shunting314/222/orig 2025-09-07T07:28:18.8760527Z * [new branch] gh/shunting314/223/base -> origin/gh/shunting314/223/base 2025-09-07T07:28:18.8761391Z * [new branch] gh/shunting314/223/head -> origin/gh/shunting314/223/head 2025-09-07T07:28:18.8762206Z * [new branch] gh/shunting314/223/orig -> origin/gh/shunting314/223/orig 2025-09-07T07:28:18.8763796Z * [new branch] gh/silverguo/1/base -> origin/gh/silverguo/1/base 2025-09-07T07:28:18.8764688Z * [new branch] gh/silverguo/1/head -> origin/gh/silverguo/1/head 2025-09-07T07:28:18.8765827Z * [new branch] gh/silverguo/2/base -> origin/gh/silverguo/2/base 2025-09-07T07:28:18.8766868Z * [new branch] gh/silverguo/2/head -> origin/gh/silverguo/2/head 2025-09-07T07:28:18.8767932Z * [new branch] gh/silverguo/3/base -> origin/gh/silverguo/3/base 2025-09-07T07:28:18.8768779Z * [new branch] gh/silverguo/3/head -> origin/gh/silverguo/3/head 2025-09-07T07:28:18.8769939Z * [new branch] gh/silverguo/4/base -> origin/gh/silverguo/4/base 2025-09-07T07:28:18.8770786Z * [new branch] gh/silverguo/4/head -> origin/gh/silverguo/4/head 2025-09-07T07:28:18.8772239Z * [new branch] gh/sinhaanhsul/1/base -> origin/gh/sinhaanhsul/1/base 2025-09-07T07:28:18.8773139Z * [new branch] gh/sinhaanhsul/1/head -> origin/gh/sinhaanhsul/1/head 2025-09-07T07:28:18.8774629Z * [new branch] gh/skarjala/17/base -> origin/gh/skarjala/17/base 2025-09-07T07:28:18.8775497Z * [new branch] gh/skarjala/17/head -> origin/gh/skarjala/17/head 2025-09-07T07:28:18.8776336Z * [new branch] gh/skarjala/17/orig -> origin/gh/skarjala/17/orig 2025-09-07T07:28:18.8777527Z * [new branch] gh/skarjala/18/base -> origin/gh/skarjala/18/base 2025-09-07T07:28:18.8778529Z * [new branch] gh/skarjala/18/head -> origin/gh/skarjala/18/head 2025-09-07T07:28:18.8779218Z * [new branch] gh/skarjala/18/orig -> origin/gh/skarjala/18/orig 2025-09-07T07:28:18.8780433Z * [new branch] gh/skarjala/19/base -> origin/gh/skarjala/19/base 2025-09-07T07:28:18.8781308Z * [new branch] gh/skarjala/19/head -> origin/gh/skarjala/19/head 2025-09-07T07:28:18.8782175Z * [new branch] gh/skarjala/19/orig -> origin/gh/skarjala/19/orig 2025-09-07T07:28:18.8783624Z * [new branch] gh/slayton58/1/base -> origin/gh/slayton58/1/base 2025-09-07T07:28:18.8784494Z * [new branch] gh/slayton58/1/head -> origin/gh/slayton58/1/head 2025-09-07T07:28:18.8785360Z * [new branch] gh/slayton58/1/orig -> origin/gh/slayton58/1/orig 2025-09-07T07:28:18.8786519Z * [new branch] gh/slayton58/2/base -> origin/gh/slayton58/2/base 2025-09-07T07:28:18.8787352Z * [new branch] gh/slayton58/2/head -> origin/gh/slayton58/2/head 2025-09-07T07:28:18.8788224Z * [new branch] gh/slayton58/2/orig -> origin/gh/slayton58/2/orig 2025-09-07T07:28:18.8789320Z * [new branch] gh/slayton58/3/base -> origin/gh/slayton58/3/base 2025-09-07T07:28:18.8790167Z * [new branch] gh/slayton58/3/head -> origin/gh/slayton58/3/head 2025-09-07T07:28:18.8791047Z * [new branch] gh/slayton58/3/orig -> origin/gh/slayton58/3/orig 2025-09-07T07:28:18.8792229Z * [new branch] gh/slayton58/4/base -> origin/gh/slayton58/4/base 2025-09-07T07:28:18.8793069Z * [new branch] gh/slayton58/4/head -> origin/gh/slayton58/4/head 2025-09-07T07:28:18.8793951Z * [new branch] gh/slayton58/4/orig -> origin/gh/slayton58/4/orig 2025-09-07T07:28:18.8795071Z * [new branch] gh/slayton58/5/base -> origin/gh/slayton58/5/base 2025-09-07T07:28:18.8795884Z * [new branch] gh/slayton58/5/head -> origin/gh/slayton58/5/head 2025-09-07T07:28:18.8796705Z * [new branch] gh/slayton58/5/orig -> origin/gh/slayton58/5/orig 2025-09-07T07:28:18.8798282Z * [new branch] gh/soulitzer/269/base -> origin/gh/soulitzer/269/base 2025-09-07T07:28:18.8799096Z * [new branch] gh/soulitzer/269/head -> origin/gh/soulitzer/269/head 2025-09-07T07:28:18.8800022Z * [new branch] gh/soulitzer/269/orig -> origin/gh/soulitzer/269/orig 2025-09-07T07:28:18.8801288Z * [new branch] gh/soulitzer/276/base -> origin/gh/soulitzer/276/base 2025-09-07T07:28:18.8802183Z * [new branch] gh/soulitzer/276/head -> origin/gh/soulitzer/276/head 2025-09-07T07:28:18.8803150Z * [new branch] gh/soulitzer/276/orig -> origin/gh/soulitzer/276/orig 2025-09-07T07:28:18.8804552Z * [new branch] gh/soulitzer/287/base -> origin/gh/soulitzer/287/base 2025-09-07T07:28:18.8805377Z * [new branch] gh/soulitzer/287/head -> origin/gh/soulitzer/287/head 2025-09-07T07:28:18.8806237Z * [new branch] gh/soulitzer/287/orig -> origin/gh/soulitzer/287/orig 2025-09-07T07:28:18.8807548Z * [new branch] gh/soulitzer/296/base -> origin/gh/soulitzer/296/base 2025-09-07T07:28:18.8808376Z * [new branch] gh/soulitzer/296/head -> origin/gh/soulitzer/296/head 2025-09-07T07:28:18.8809229Z * [new branch] gh/soulitzer/296/orig -> origin/gh/soulitzer/296/orig 2025-09-07T07:28:18.8810435Z * [new branch] gh/soulitzer/299/base -> origin/gh/soulitzer/299/base 2025-09-07T07:28:18.8811374Z * [new branch] gh/soulitzer/299/head -> origin/gh/soulitzer/299/head 2025-09-07T07:28:18.8812166Z * [new branch] gh/soulitzer/299/orig -> origin/gh/soulitzer/299/orig 2025-09-07T07:28:18.8813479Z * [new branch] gh/soulitzer/300/base -> origin/gh/soulitzer/300/base 2025-09-07T07:28:18.8814269Z * [new branch] gh/soulitzer/300/head -> origin/gh/soulitzer/300/head 2025-09-07T07:28:18.8815193Z * [new branch] gh/soulitzer/300/orig -> origin/gh/soulitzer/300/orig 2025-09-07T07:28:18.8816493Z * [new branch] gh/soulitzer/301/base -> origin/gh/soulitzer/301/base 2025-09-07T07:28:18.8817352Z * [new branch] gh/soulitzer/301/head -> origin/gh/soulitzer/301/head 2025-09-07T07:28:18.8818234Z * [new branch] gh/soulitzer/301/orig -> origin/gh/soulitzer/301/orig 2025-09-07T07:28:18.8819418Z * [new branch] gh/soulitzer/313/base -> origin/gh/soulitzer/313/base 2025-09-07T07:28:18.8820273Z * [new branch] gh/soulitzer/313/head -> origin/gh/soulitzer/313/head 2025-09-07T07:28:18.8821207Z * [new branch] gh/soulitzer/313/orig -> origin/gh/soulitzer/313/orig 2025-09-07T07:28:18.8822428Z * [new branch] gh/soulitzer/319/base -> origin/gh/soulitzer/319/base 2025-09-07T07:28:18.8823246Z * [new branch] gh/soulitzer/319/head -> origin/gh/soulitzer/319/head 2025-09-07T07:28:18.8824046Z * [new branch] gh/soulitzer/319/orig -> origin/gh/soulitzer/319/orig 2025-09-07T07:28:18.8825360Z * [new branch] gh/soulitzer/320/base -> origin/gh/soulitzer/320/base 2025-09-07T07:28:18.8826160Z * [new branch] gh/soulitzer/320/head -> origin/gh/soulitzer/320/head 2025-09-07T07:28:18.8826998Z * [new branch] gh/soulitzer/320/orig -> origin/gh/soulitzer/320/orig 2025-09-07T07:28:18.8828315Z * [new branch] gh/soulitzer/336/base -> origin/gh/soulitzer/336/base 2025-09-07T07:28:18.8829086Z * [new branch] gh/soulitzer/336/head -> origin/gh/soulitzer/336/head 2025-09-07T07:28:18.8829934Z * [new branch] gh/soulitzer/336/orig -> origin/gh/soulitzer/336/orig 2025-09-07T07:28:18.8831173Z * [new branch] gh/soulitzer/347/base -> origin/gh/soulitzer/347/base 2025-09-07T07:28:18.8831973Z * [new branch] gh/soulitzer/347/head -> origin/gh/soulitzer/347/head 2025-09-07T07:28:18.8832809Z * [new branch] gh/soulitzer/347/orig -> origin/gh/soulitzer/347/orig 2025-09-07T07:28:18.8834195Z * [new branch] gh/soulitzer/349/base -> origin/gh/soulitzer/349/base 2025-09-07T07:28:18.8835048Z * [new branch] gh/soulitzer/349/head -> origin/gh/soulitzer/349/head 2025-09-07T07:28:18.8835933Z * [new branch] gh/soulitzer/349/orig -> origin/gh/soulitzer/349/orig 2025-09-07T07:28:18.8837023Z * [new branch] gh/soulitzer/350/base -> origin/gh/soulitzer/350/base 2025-09-07T07:28:18.8837826Z * [new branch] gh/soulitzer/350/head -> origin/gh/soulitzer/350/head 2025-09-07T07:28:18.8838750Z * [new branch] gh/soulitzer/350/orig -> origin/gh/soulitzer/350/orig 2025-09-07T07:28:18.8839968Z * [new branch] gh/soulitzer/351/base -> origin/gh/soulitzer/351/base 2025-09-07T07:28:18.8840869Z * [new branch] gh/soulitzer/351/head -> origin/gh/soulitzer/351/head 2025-09-07T07:28:18.8841681Z * [new branch] gh/soulitzer/351/orig -> origin/gh/soulitzer/351/orig 2025-09-07T07:28:18.8842844Z * [new branch] gh/soulitzer/353/base -> origin/gh/soulitzer/353/base 2025-09-07T07:28:18.8843773Z * [new branch] gh/soulitzer/353/head -> origin/gh/soulitzer/353/head 2025-09-07T07:28:18.8844569Z * [new branch] gh/soulitzer/353/orig -> origin/gh/soulitzer/353/orig 2025-09-07T07:28:18.8846349Z * [new branch] gh/soulitzer/358/base -> origin/gh/soulitzer/358/base 2025-09-07T07:28:18.8847310Z * [new branch] gh/soulitzer/358/head -> origin/gh/soulitzer/358/head 2025-09-07T07:28:18.8848287Z * [new branch] gh/soulitzer/358/orig -> origin/gh/soulitzer/358/orig 2025-09-07T07:28:18.8849754Z * [new branch] gh/soulitzer/359/base -> origin/gh/soulitzer/359/base 2025-09-07T07:28:18.8850673Z * [new branch] gh/soulitzer/359/head -> origin/gh/soulitzer/359/head 2025-09-07T07:28:18.8851589Z * [new branch] gh/soulitzer/359/orig -> origin/gh/soulitzer/359/orig 2025-09-07T07:28:18.8852883Z * [new branch] gh/soulitzer/362/base -> origin/gh/soulitzer/362/base 2025-09-07T07:28:18.8853708Z * [new branch] gh/soulitzer/362/head -> origin/gh/soulitzer/362/head 2025-09-07T07:28:18.8854570Z * [new branch] gh/soulitzer/362/orig -> origin/gh/soulitzer/362/orig 2025-09-07T07:28:18.8855728Z * [new branch] gh/soulitzer/372/base -> origin/gh/soulitzer/372/base 2025-09-07T07:28:18.8856629Z * [new branch] gh/soulitzer/372/head -> origin/gh/soulitzer/372/head 2025-09-07T07:28:18.8857468Z * [new branch] gh/soulitzer/372/orig -> origin/gh/soulitzer/372/orig 2025-09-07T07:28:18.8858662Z * [new branch] gh/soulitzer/373/base -> origin/gh/soulitzer/373/base 2025-09-07T07:28:18.8859569Z * [new branch] gh/soulitzer/373/head -> origin/gh/soulitzer/373/head 2025-09-07T07:28:18.8860363Z * [new branch] gh/soulitzer/373/orig -> origin/gh/soulitzer/373/orig 2025-09-07T07:28:18.8861595Z * [new branch] gh/soulitzer/374/base -> origin/gh/soulitzer/374/base 2025-09-07T07:28:18.8862470Z * [new branch] gh/soulitzer/374/head -> origin/gh/soulitzer/374/head 2025-09-07T07:28:18.8863311Z * [new branch] gh/soulitzer/374/orig -> origin/gh/soulitzer/374/orig 2025-09-07T07:28:18.8864524Z * [new branch] gh/soulitzer/375/base -> origin/gh/soulitzer/375/base 2025-09-07T07:28:18.8865327Z * [new branch] gh/soulitzer/375/head -> origin/gh/soulitzer/375/head 2025-09-07T07:28:18.8869205Z * [new branch] gh/soulitzer/375/orig -> origin/gh/soulitzer/375/orig 2025-09-07T07:28:18.8870396Z * [new branch] gh/soulitzer/376/base -> origin/gh/soulitzer/376/base 2025-09-07T07:28:18.8871311Z * [new branch] gh/soulitzer/376/head -> origin/gh/soulitzer/376/head 2025-09-07T07:28:18.8872109Z * [new branch] gh/soulitzer/376/orig -> origin/gh/soulitzer/376/orig 2025-09-07T07:28:18.8873354Z * [new branch] gh/soulitzer/377/base -> origin/gh/soulitzer/377/base 2025-09-07T07:28:18.8874191Z * [new branch] gh/soulitzer/377/head -> origin/gh/soulitzer/377/head 2025-09-07T07:28:18.8874999Z * [new branch] gh/soulitzer/377/orig -> origin/gh/soulitzer/377/orig 2025-09-07T07:28:18.8876366Z * [new branch] gh/soulitzer/378/base -> origin/gh/soulitzer/378/base 2025-09-07T07:28:18.8877250Z * [new branch] gh/soulitzer/378/head -> origin/gh/soulitzer/378/head 2025-09-07T07:28:18.8878222Z * [new branch] gh/soulitzer/378/orig -> origin/gh/soulitzer/378/orig 2025-09-07T07:28:18.8879409Z * [new branch] gh/soulitzer/379/base -> origin/gh/soulitzer/379/base 2025-09-07T07:28:18.8880299Z * [new branch] gh/soulitzer/379/head -> origin/gh/soulitzer/379/head 2025-09-07T07:28:18.8881138Z * [new branch] gh/soulitzer/379/orig -> origin/gh/soulitzer/379/orig 2025-09-07T07:28:18.8882723Z * [new branch] gh/swolchok/728/next -> origin/gh/swolchok/728/next 2025-09-07T07:28:18.8884243Z * [new branch] gh/swolchok/767/base -> origin/gh/swolchok/767/base 2025-09-07T07:28:18.8885313Z * [new branch] gh/swolchok/767/head -> origin/gh/swolchok/767/head 2025-09-07T07:28:18.8886515Z * [new branch] gh/swolchok/767/orig -> origin/gh/swolchok/767/orig 2025-09-07T07:28:18.8887596Z * [new branch] gh/swolchok/768/base -> origin/gh/swolchok/768/base 2025-09-07T07:28:18.8888495Z * [new branch] gh/swolchok/768/head -> origin/gh/swolchok/768/head 2025-09-07T07:28:18.8889442Z * [new branch] gh/swolchok/768/orig -> origin/gh/swolchok/768/orig 2025-09-07T07:28:18.8890888Z * [new branch] gh/swolchok/769/base -> origin/gh/swolchok/769/base 2025-09-07T07:28:18.8891827Z * [new branch] gh/swolchok/769/head -> origin/gh/swolchok/769/head 2025-09-07T07:28:18.8892796Z * [new branch] gh/swolchok/769/orig -> origin/gh/swolchok/769/orig 2025-09-07T07:28:18.8894008Z * [new branch] gh/swolchok/771/base -> origin/gh/swolchok/771/base 2025-09-07T07:28:18.8894920Z * [new branch] gh/swolchok/771/head -> origin/gh/swolchok/771/head 2025-09-07T07:28:18.8895805Z * [new branch] gh/swolchok/771/orig -> origin/gh/swolchok/771/orig 2025-09-07T07:28:18.8896953Z * [new branch] gh/swolchok/772/base -> origin/gh/swolchok/772/base 2025-09-07T07:28:18.8897873Z * [new branch] gh/swolchok/772/head -> origin/gh/swolchok/772/head 2025-09-07T07:28:18.8898806Z * [new branch] gh/swolchok/772/orig -> origin/gh/swolchok/772/orig 2025-09-07T07:28:18.8900040Z * [new branch] gh/swolchok/773/base -> origin/gh/swolchok/773/base 2025-09-07T07:28:18.8900931Z * [new branch] gh/swolchok/773/head -> origin/gh/swolchok/773/head 2025-09-07T07:28:18.8901892Z * [new branch] gh/swolchok/773/orig -> origin/gh/swolchok/773/orig 2025-09-07T07:28:18.8903129Z * [new branch] gh/swolchok/786/base -> origin/gh/swolchok/786/base 2025-09-07T07:28:18.8903963Z * [new branch] gh/swolchok/786/head -> origin/gh/swolchok/786/head 2025-09-07T07:28:18.8904845Z * [new branch] gh/swolchok/786/orig -> origin/gh/swolchok/786/orig 2025-09-07T07:28:18.8906009Z * [new branch] gh/swolchok/787/base -> origin/gh/swolchok/787/base 2025-09-07T07:28:18.8907121Z * [new branch] gh/swolchok/787/head -> origin/gh/swolchok/787/head 2025-09-07T07:28:18.8908034Z * [new branch] gh/swolchok/787/orig -> origin/gh/swolchok/787/orig 2025-09-07T07:28:18.8909222Z * [new branch] gh/swolchok/788/base -> origin/gh/swolchok/788/base 2025-09-07T07:28:18.8910108Z * [new branch] gh/swolchok/788/head -> origin/gh/swolchok/788/head 2025-09-07T07:28:18.8911227Z * [new branch] gh/swolchok/788/orig -> origin/gh/swolchok/788/orig 2025-09-07T07:28:18.8912413Z * [new branch] gh/swolchok/789/base -> origin/gh/swolchok/789/base 2025-09-07T07:28:18.8913249Z * [new branch] gh/swolchok/789/head -> origin/gh/swolchok/789/head 2025-09-07T07:28:18.8914172Z * [new branch] gh/swolchok/789/orig -> origin/gh/swolchok/789/orig 2025-09-07T07:28:18.8915552Z * [new branch] gh/swolchok/790/base -> origin/gh/swolchok/790/base 2025-09-07T07:28:18.8916490Z * [new branch] gh/swolchok/790/head -> origin/gh/swolchok/790/head 2025-09-07T07:28:18.8917343Z * [new branch] gh/swolchok/790/orig -> origin/gh/swolchok/790/orig 2025-09-07T07:28:18.8918578Z * [new branch] gh/swolchok/791/base -> origin/gh/swolchok/791/base 2025-09-07T07:28:18.8919380Z * [new branch] gh/swolchok/791/head -> origin/gh/swolchok/791/head 2025-09-07T07:28:18.8920234Z * [new branch] gh/swolchok/791/orig -> origin/gh/swolchok/791/orig 2025-09-07T07:28:18.8921499Z * [new branch] gh/swolchok/792/base -> origin/gh/swolchok/792/base 2025-09-07T07:28:18.8922355Z * [new branch] gh/swolchok/792/head -> origin/gh/swolchok/792/head 2025-09-07T07:28:18.8923139Z * [new branch] gh/swolchok/792/orig -> origin/gh/swolchok/792/orig 2025-09-07T07:28:18.8924343Z * [new branch] gh/swolchok/793/base -> origin/gh/swolchok/793/base 2025-09-07T07:28:18.8925146Z * [new branch] gh/swolchok/793/head -> origin/gh/swolchok/793/head 2025-09-07T07:28:18.8925997Z * [new branch] gh/swolchok/793/orig -> origin/gh/swolchok/793/orig 2025-09-07T07:28:18.8927335Z * [new branch] gh/swolchok/794/base -> origin/gh/swolchok/794/base 2025-09-07T07:28:18.8928157Z * [new branch] gh/swolchok/794/head -> origin/gh/swolchok/794/head 2025-09-07T07:28:18.8928967Z * [new branch] gh/swolchok/794/orig -> origin/gh/swolchok/794/orig 2025-09-07T07:28:18.8930547Z * [new branch] gh/swolchok/795/base -> origin/gh/swolchok/795/base 2025-09-07T07:28:18.8931511Z * [new branch] gh/swolchok/795/head -> origin/gh/swolchok/795/head 2025-09-07T07:28:18.8932368Z * [new branch] gh/swolchok/795/orig -> origin/gh/swolchok/795/orig 2025-09-07T07:28:18.8933566Z * [new branch] gh/swolchok/796/base -> origin/gh/swolchok/796/base 2025-09-07T07:28:18.8934614Z * [new branch] gh/swolchok/796/head -> origin/gh/swolchok/796/head 2025-09-07T07:28:18.8935445Z * [new branch] gh/swolchok/796/orig -> origin/gh/swolchok/796/orig 2025-09-07T07:28:18.8936821Z * [new branch] gh/swolchok/797/base -> origin/gh/swolchok/797/base 2025-09-07T07:28:18.8937674Z * [new branch] gh/swolchok/797/head -> origin/gh/swolchok/797/head 2025-09-07T07:28:18.8938593Z * [new branch] gh/swolchok/797/orig -> origin/gh/swolchok/797/orig 2025-09-07T07:28:18.8939913Z * [new branch] gh/swolchok/798/base -> origin/gh/swolchok/798/base 2025-09-07T07:28:18.8940761Z * [new branch] gh/swolchok/798/head -> origin/gh/swolchok/798/head 2025-09-07T07:28:18.8941648Z * [new branch] gh/swolchok/798/orig -> origin/gh/swolchok/798/orig 2025-09-07T07:28:18.8943016Z * [new branch] gh/swolchok/799/base -> origin/gh/swolchok/799/base 2025-09-07T07:28:18.8943795Z * [new branch] gh/swolchok/799/head -> origin/gh/swolchok/799/head 2025-09-07T07:28:18.8944787Z * [new branch] gh/swolchok/799/orig -> origin/gh/swolchok/799/orig 2025-09-07T07:28:18.8946197Z * [new branch] gh/swolchok/800/base -> origin/gh/swolchok/800/base 2025-09-07T07:28:18.8947008Z * [new branch] gh/swolchok/800/head -> origin/gh/swolchok/800/head 2025-09-07T07:28:18.8947963Z * [new branch] gh/swolchok/800/orig -> origin/gh/swolchok/800/orig 2025-09-07T07:28:18.8949255Z * [new branch] gh/swolchok/801/base -> origin/gh/swolchok/801/base 2025-09-07T07:28:18.8950068Z * [new branch] gh/swolchok/801/head -> origin/gh/swolchok/801/head 2025-09-07T07:28:18.8951032Z * [new branch] gh/swolchok/801/orig -> origin/gh/swolchok/801/orig 2025-09-07T07:28:18.8952352Z * [new branch] gh/swolchok/802/base -> origin/gh/swolchok/802/base 2025-09-07T07:28:18.8953174Z * [new branch] gh/swolchok/802/head -> origin/gh/swolchok/802/head 2025-09-07T07:28:18.8954079Z * [new branch] gh/swolchok/802/orig -> origin/gh/swolchok/802/orig 2025-09-07T07:28:18.8955277Z * [new branch] gh/swolchok/803/base -> origin/gh/swolchok/803/base 2025-09-07T07:28:18.8956129Z * [new branch] gh/swolchok/803/head -> origin/gh/swolchok/803/head 2025-09-07T07:28:18.8957055Z * [new branch] gh/swolchok/803/orig -> origin/gh/swolchok/803/orig 2025-09-07T07:28:18.8958494Z * [new branch] gh/swolchok/804/base -> origin/gh/swolchok/804/base 2025-09-07T07:28:18.8959192Z * [new branch] gh/swolchok/804/head -> origin/gh/swolchok/804/head 2025-09-07T07:28:18.8960164Z * [new branch] gh/swolchok/804/orig -> origin/gh/swolchok/804/orig 2025-09-07T07:28:18.8961377Z * [new branch] gh/swolchok/805/base -> origin/gh/swolchok/805/base 2025-09-07T07:28:18.8962287Z * [new branch] gh/swolchok/805/head -> origin/gh/swolchok/805/head 2025-09-07T07:28:18.8963163Z * [new branch] gh/swolchok/805/orig -> origin/gh/swolchok/805/orig 2025-09-07T07:28:18.8964358Z * [new branch] gh/swolchok/806/base -> origin/gh/swolchok/806/base 2025-09-07T07:28:18.8965238Z * [new branch] gh/swolchok/806/head -> origin/gh/swolchok/806/head 2025-09-07T07:28:18.8966258Z * [new branch] gh/swolchok/806/orig -> origin/gh/swolchok/806/orig 2025-09-07T07:28:18.8967898Z * [new branch] gh/swolchok/807/base -> origin/gh/swolchok/807/base 2025-09-07T07:28:18.8968705Z * [new branch] gh/swolchok/807/head -> origin/gh/swolchok/807/head 2025-09-07T07:28:18.8969730Z * [new branch] gh/swolchok/807/orig -> origin/gh/swolchok/807/orig 2025-09-07T07:28:18.8971059Z * [new branch] gh/swolchok/808/base -> origin/gh/swolchok/808/base 2025-09-07T07:28:18.8971939Z * [new branch] gh/swolchok/808/head -> origin/gh/swolchok/808/head 2025-09-07T07:28:18.8972761Z * [new branch] gh/swolchok/808/orig -> origin/gh/swolchok/808/orig 2025-09-07T07:28:18.8973944Z * [new branch] gh/swolchok/809/base -> origin/gh/swolchok/809/base 2025-09-07T07:28:18.8974844Z * [new branch] gh/swolchok/809/head -> origin/gh/swolchok/809/head 2025-09-07T07:28:18.8975728Z * [new branch] gh/swolchok/809/orig -> origin/gh/swolchok/809/orig 2025-09-07T07:28:18.8977145Z * [new branch] gh/swolchok/810/base -> origin/gh/swolchok/810/base 2025-09-07T07:28:18.8977990Z * [new branch] gh/swolchok/810/head -> origin/gh/swolchok/810/head 2025-09-07T07:28:18.8978801Z * [new branch] gh/swolchok/810/orig -> origin/gh/swolchok/810/orig 2025-09-07T07:28:18.8980006Z * [new branch] gh/swolchok/811/base -> origin/gh/swolchok/811/base 2025-09-07T07:28:18.8980962Z * [new branch] gh/swolchok/811/head -> origin/gh/swolchok/811/head 2025-09-07T07:28:18.8981889Z * [new branch] gh/swolchok/811/orig -> origin/gh/swolchok/811/orig 2025-09-07T07:28:18.8983194Z * [new branch] gh/swolchok/812/base -> origin/gh/swolchok/812/base 2025-09-07T07:28:18.8984028Z * [new branch] gh/swolchok/812/head -> origin/gh/swolchok/812/head 2025-09-07T07:28:18.8984902Z * [new branch] gh/swolchok/812/orig -> origin/gh/swolchok/812/orig 2025-09-07T07:28:18.8986267Z * [new branch] gh/swolchok/813/base -> origin/gh/swolchok/813/base 2025-09-07T07:28:18.8987073Z * [new branch] gh/swolchok/813/head -> origin/gh/swolchok/813/head 2025-09-07T07:28:18.8987992Z * [new branch] gh/swolchok/813/orig -> origin/gh/swolchok/813/orig 2025-09-07T07:28:18.8989307Z * [new branch] gh/swolchok/814/base -> origin/gh/swolchok/814/base 2025-09-07T07:28:18.8990116Z * [new branch] gh/swolchok/814/head -> origin/gh/swolchok/814/head 2025-09-07T07:28:18.8990936Z * [new branch] gh/swolchok/814/orig -> origin/gh/swolchok/814/orig 2025-09-07T07:28:18.8992204Z * [new branch] gh/swolchok/815/base -> origin/gh/swolchok/815/base 2025-09-07T07:28:18.8993036Z * [new branch] gh/swolchok/815/head -> origin/gh/swolchok/815/head 2025-09-07T07:28:18.8994068Z * [new branch] gh/swolchok/815/orig -> origin/gh/swolchok/815/orig 2025-09-07T07:28:18.8995093Z * [new branch] gh/swolchok/816/base -> origin/gh/swolchok/816/base 2025-09-07T07:28:18.8996027Z * [new branch] gh/swolchok/816/head -> origin/gh/swolchok/816/head 2025-09-07T07:28:18.8996891Z * [new branch] gh/swolchok/816/orig -> origin/gh/swolchok/816/orig 2025-09-07T07:28:18.8998139Z * [new branch] gh/swolchok/817/base -> origin/gh/swolchok/817/base 2025-09-07T07:28:18.8998999Z * [new branch] gh/swolchok/817/head -> origin/gh/swolchok/817/head 2025-09-07T07:28:18.8999815Z * [new branch] gh/swolchok/817/orig -> origin/gh/swolchok/817/orig 2025-09-07T07:28:18.9001124Z * [new branch] gh/swolchok/818/base -> origin/gh/swolchok/818/base 2025-09-07T07:28:18.9001951Z * [new branch] gh/swolchok/818/head -> origin/gh/swolchok/818/head 2025-09-07T07:28:18.9002801Z * [new branch] gh/swolchok/818/orig -> origin/gh/swolchok/818/orig 2025-09-07T07:28:18.9004157Z * [new branch] gh/swolchok/819/base -> origin/gh/swolchok/819/base 2025-09-07T07:28:18.9004968Z * [new branch] gh/swolchok/819/head -> origin/gh/swolchok/819/head 2025-09-07T07:28:18.9005817Z * [new branch] gh/swolchok/819/orig -> origin/gh/swolchok/819/orig 2025-09-07T07:28:18.9007137Z * [new branch] gh/swolchok/820/base -> origin/gh/swolchok/820/base 2025-09-07T07:28:18.9007906Z * [new branch] gh/swolchok/820/head -> origin/gh/swolchok/820/head 2025-09-07T07:28:18.9008797Z * [new branch] gh/swolchok/820/orig -> origin/gh/swolchok/820/orig 2025-09-07T07:28:18.9010006Z * [new branch] gh/swolchok/821/base -> origin/gh/swolchok/821/base 2025-09-07T07:28:18.9010801Z * [new branch] gh/swolchok/821/head -> origin/gh/swolchok/821/head 2025-09-07T07:28:18.9011674Z * [new branch] gh/swolchok/821/orig -> origin/gh/swolchok/821/orig 2025-09-07T07:28:18.9013068Z * [new branch] gh/swolchok/822/base -> origin/gh/swolchok/822/base 2025-09-07T07:28:18.9013919Z * [new branch] gh/swolchok/822/head -> origin/gh/swolchok/822/head 2025-09-07T07:28:18.9014807Z * [new branch] gh/swolchok/822/orig -> origin/gh/swolchok/822/orig 2025-09-07T07:28:18.9016075Z * [new branch] gh/swolchok/823/base -> origin/gh/swolchok/823/base 2025-09-07T07:28:18.9016893Z * [new branch] gh/swolchok/823/head -> origin/gh/swolchok/823/head 2025-09-07T07:28:18.9017710Z * [new branch] gh/swolchok/823/orig -> origin/gh/swolchok/823/orig 2025-09-07T07:28:18.9018911Z * [new branch] gh/swolchok/824/base -> origin/gh/swolchok/824/base 2025-09-07T07:28:18.9019741Z * [new branch] gh/swolchok/824/head -> origin/gh/swolchok/824/head 2025-09-07T07:28:18.9020590Z * [new branch] gh/swolchok/824/orig -> origin/gh/swolchok/824/orig 2025-09-07T07:28:18.9021762Z * [new branch] gh/swolchok/825/base -> origin/gh/swolchok/825/base 2025-09-07T07:28:18.9022631Z * [new branch] gh/swolchok/825/head -> origin/gh/swolchok/825/head 2025-09-07T07:28:18.9023545Z * [new branch] gh/swolchok/825/orig -> origin/gh/swolchok/825/orig 2025-09-07T07:28:18.9024927Z * [new branch] gh/swolchok/826/base -> origin/gh/swolchok/826/base 2025-09-07T07:28:18.9025726Z * [new branch] gh/swolchok/826/head -> origin/gh/swolchok/826/head 2025-09-07T07:28:18.9026522Z * [new branch] gh/swolchok/826/orig -> origin/gh/swolchok/826/orig 2025-09-07T07:28:18.9027782Z * [new branch] gh/swolchok/827/base -> origin/gh/swolchok/827/base 2025-09-07T07:28:18.9028668Z * [new branch] gh/swolchok/827/head -> origin/gh/swolchok/827/head 2025-09-07T07:28:18.9029390Z * [new branch] gh/swolchok/827/orig -> origin/gh/swolchok/827/orig 2025-09-07T07:28:18.9030717Z * [new branch] gh/swolchok/828/base -> origin/gh/swolchok/828/base 2025-09-07T07:28:18.9031518Z * [new branch] gh/swolchok/828/head -> origin/gh/swolchok/828/head 2025-09-07T07:28:18.9032341Z * [new branch] gh/swolchok/828/orig -> origin/gh/swolchok/828/orig 2025-09-07T07:28:18.9033493Z * [new branch] gh/swolchok/829/base -> origin/gh/swolchok/829/base 2025-09-07T07:28:18.9034360Z * [new branch] gh/swolchok/829/head -> origin/gh/swolchok/829/head 2025-09-07T07:28:18.9035230Z * [new branch] gh/swolchok/829/orig -> origin/gh/swolchok/829/orig 2025-09-07T07:28:18.9036620Z * [new branch] gh/swolchok/830/base -> origin/gh/swolchok/830/base 2025-09-07T07:28:18.9037419Z * [new branch] gh/swolchok/830/head -> origin/gh/swolchok/830/head 2025-09-07T07:28:18.9038176Z * [new branch] gh/swolchok/830/orig -> origin/gh/swolchok/830/orig 2025-09-07T07:28:18.9039279Z * [new branch] gh/swolchok/831/base -> origin/gh/swolchok/831/base 2025-09-07T07:28:18.9040161Z * [new branch] gh/swolchok/831/head -> origin/gh/swolchok/831/head 2025-09-07T07:28:18.9040953Z * [new branch] gh/swolchok/831/orig -> origin/gh/swolchok/831/orig 2025-09-07T07:28:18.9042086Z * [new branch] gh/swolchok/832/base -> origin/gh/swolchok/832/base 2025-09-07T07:28:18.9043015Z * [new branch] gh/swolchok/832/head -> origin/gh/swolchok/832/head 2025-09-07T07:28:18.9043790Z * [new branch] gh/swolchok/832/orig -> origin/gh/swolchok/832/orig 2025-09-07T07:28:18.9045278Z * [new branch] gh/syed-ahmed/3/base -> origin/gh/syed-ahmed/3/base 2025-09-07T07:28:18.9046198Z * [new branch] gh/syed-ahmed/3/head -> origin/gh/syed-ahmed/3/head 2025-09-07T07:28:18.9047082Z * [new branch] gh/syed-ahmed/3/orig -> origin/gh/syed-ahmed/3/orig 2025-09-07T07:28:18.9048244Z * [new branch] gh/syed-ahmed/4/base -> origin/gh/syed-ahmed/4/base 2025-09-07T07:28:18.9049311Z * [new branch] gh/syed-ahmed/4/head -> origin/gh/syed-ahmed/4/head 2025-09-07T07:28:18.9050201Z * [new branch] gh/syed-ahmed/4/orig -> origin/gh/syed-ahmed/4/orig 2025-09-07T07:28:18.9051366Z * [new branch] gh/syed-ahmed/5/base -> origin/gh/syed-ahmed/5/base 2025-09-07T07:28:18.9052173Z * [new branch] gh/syed-ahmed/5/head -> origin/gh/syed-ahmed/5/head 2025-09-07T07:28:18.9053052Z * [new branch] gh/syed-ahmed/5/orig -> origin/gh/syed-ahmed/5/orig 2025-09-07T07:28:18.9054580Z * [new branch] gh/teja-rao/4/base -> origin/gh/teja-rao/4/base 2025-09-07T07:28:18.9055515Z * [new branch] gh/teja-rao/4/head -> origin/gh/teja-rao/4/head 2025-09-07T07:28:18.9056426Z * [new branch] gh/teja-rao/4/orig -> origin/gh/teja-rao/4/orig 2025-09-07T07:28:18.9057890Z * [new branch] gh/tianyu-l/2/base -> origin/gh/tianyu-l/2/base 2025-09-07T07:28:18.9058801Z * [new branch] gh/tianyu-l/2/head -> origin/gh/tianyu-l/2/head 2025-09-07T07:28:18.9059618Z * [new branch] gh/tianyu-l/2/orig -> origin/gh/tianyu-l/2/orig 2025-09-07T07:28:18.9060766Z * [new branch] gh/tianyu-l/3/base -> origin/gh/tianyu-l/3/base 2025-09-07T07:28:18.9061832Z * [new branch] gh/tianyu-l/3/head -> origin/gh/tianyu-l/3/head 2025-09-07T07:28:18.9062736Z * [new branch] gh/tianyu-l/3/orig -> origin/gh/tianyu-l/3/orig 2025-09-07T07:28:18.9063901Z * [new branch] gh/tianyu-l/4/base -> origin/gh/tianyu-l/4/base 2025-09-07T07:28:18.9064826Z * [new branch] gh/tianyu-l/4/head -> origin/gh/tianyu-l/4/head 2025-09-07T07:28:18.9065659Z * [new branch] gh/tianyu-l/4/orig -> origin/gh/tianyu-l/4/orig 2025-09-07T07:28:18.9067252Z * [new branch] gh/tugsbayasgalan/1/base -> origin/gh/tugsbayasgalan/1/base 2025-09-07T07:28:18.9068084Z * [new branch] gh/tugsbayasgalan/1/head -> origin/gh/tugsbayasgalan/1/head 2025-09-07T07:28:18.9069072Z * [new branch] gh/tugsbayasgalan/1/orig -> origin/gh/tugsbayasgalan/1/orig 2025-09-07T07:28:18.9070561Z * [new branch] gh/tugsbayasgalan/10/base -> origin/gh/tugsbayasgalan/10/base 2025-09-07T07:28:18.9071470Z * [new branch] gh/tugsbayasgalan/10/head -> origin/gh/tugsbayasgalan/10/head 2025-09-07T07:28:18.9072344Z * [new branch] gh/tugsbayasgalan/10/orig -> origin/gh/tugsbayasgalan/10/orig 2025-09-07T07:28:18.9073422Z * [new branch] gh/tugsbayasgalan/11/base -> origin/gh/tugsbayasgalan/11/base 2025-09-07T07:28:18.9074323Z * [new branch] gh/tugsbayasgalan/11/head -> origin/gh/tugsbayasgalan/11/head 2025-09-07T07:28:18.9075112Z * [new branch] gh/tugsbayasgalan/11/orig -> origin/gh/tugsbayasgalan/11/orig 2025-09-07T07:28:18.9076410Z * [new branch] gh/tugsbayasgalan/12/base -> origin/gh/tugsbayasgalan/12/base 2025-09-07T07:28:18.9077249Z * [new branch] gh/tugsbayasgalan/12/head -> origin/gh/tugsbayasgalan/12/head 2025-09-07T07:28:18.9078080Z * [new branch] gh/tugsbayasgalan/12/orig -> origin/gh/tugsbayasgalan/12/orig 2025-09-07T07:28:18.9079229Z * [new branch] gh/tugsbayasgalan/13/base -> origin/gh/tugsbayasgalan/13/base 2025-09-07T07:28:18.9080092Z * [new branch] gh/tugsbayasgalan/13/head -> origin/gh/tugsbayasgalan/13/head 2025-09-07T07:28:18.9080939Z * [new branch] gh/tugsbayasgalan/13/orig -> origin/gh/tugsbayasgalan/13/orig 2025-09-07T07:28:18.9082263Z * [new branch] gh/tugsbayasgalan/14/base -> origin/gh/tugsbayasgalan/14/base 2025-09-07T07:28:18.9083075Z * [new branch] gh/tugsbayasgalan/14/head -> origin/gh/tugsbayasgalan/14/head 2025-09-07T07:28:18.9083949Z * [new branch] gh/tugsbayasgalan/14/orig -> origin/gh/tugsbayasgalan/14/orig 2025-09-07T07:28:18.9085299Z * [new branch] gh/tugsbayasgalan/15/base -> origin/gh/tugsbayasgalan/15/base 2025-09-07T07:28:18.9086081Z * [new branch] gh/tugsbayasgalan/15/head -> origin/gh/tugsbayasgalan/15/head 2025-09-07T07:28:18.9086866Z * [new branch] gh/tugsbayasgalan/15/orig -> origin/gh/tugsbayasgalan/15/orig 2025-09-07T07:28:18.9088118Z * [new branch] gh/tugsbayasgalan/2/base -> origin/gh/tugsbayasgalan/2/base 2025-09-07T07:28:18.9088927Z * [new branch] gh/tugsbayasgalan/2/head -> origin/gh/tugsbayasgalan/2/head 2025-09-07T07:28:18.9089783Z * [new branch] gh/tugsbayasgalan/2/orig -> origin/gh/tugsbayasgalan/2/orig 2025-09-07T07:28:18.9090835Z * [new branch] gh/tugsbayasgalan/3/base -> origin/gh/tugsbayasgalan/3/base 2025-09-07T07:28:18.9091840Z * [new branch] gh/tugsbayasgalan/3/head -> origin/gh/tugsbayasgalan/3/head 2025-09-07T07:28:18.9092667Z * [new branch] gh/tugsbayasgalan/3/orig -> origin/gh/tugsbayasgalan/3/orig 2025-09-07T07:28:18.9093915Z * [new branch] gh/tugsbayasgalan/4/base -> origin/gh/tugsbayasgalan/4/base 2025-09-07T07:28:18.9094960Z * [new branch] gh/tugsbayasgalan/4/head -> origin/gh/tugsbayasgalan/4/head 2025-09-07T07:28:18.9095809Z * [new branch] gh/tugsbayasgalan/4/orig -> origin/gh/tugsbayasgalan/4/orig 2025-09-07T07:28:18.9097095Z * [new branch] gh/tugsbayasgalan/5/base -> origin/gh/tugsbayasgalan/5/base 2025-09-07T07:28:18.9098074Z * [new branch] gh/tugsbayasgalan/5/head -> origin/gh/tugsbayasgalan/5/head 2025-09-07T07:28:18.9098915Z * [new branch] gh/tugsbayasgalan/5/orig -> origin/gh/tugsbayasgalan/5/orig 2025-09-07T07:28:18.9100311Z * [new branch] gh/tugsbayasgalan/6/base -> origin/gh/tugsbayasgalan/6/base 2025-09-07T07:28:18.9101114Z * [new branch] gh/tugsbayasgalan/6/head -> origin/gh/tugsbayasgalan/6/head 2025-09-07T07:28:18.9101977Z * [new branch] gh/tugsbayasgalan/6/orig -> origin/gh/tugsbayasgalan/6/orig 2025-09-07T07:28:18.9103434Z * [new branch] gh/tugsbayasgalan/7/base -> origin/gh/tugsbayasgalan/7/base 2025-09-07T07:28:18.9104339Z * [new branch] gh/tugsbayasgalan/7/head -> origin/gh/tugsbayasgalan/7/head 2025-09-07T07:28:18.9105183Z * [new branch] gh/tugsbayasgalan/7/orig -> origin/gh/tugsbayasgalan/7/orig 2025-09-07T07:28:18.9106438Z * [new branch] gh/tugsbayasgalan/8/base -> origin/gh/tugsbayasgalan/8/base 2025-09-07T07:28:18.9107252Z * [new branch] gh/tugsbayasgalan/8/head -> origin/gh/tugsbayasgalan/8/head 2025-09-07T07:28:18.9108098Z * [new branch] gh/tugsbayasgalan/8/orig -> origin/gh/tugsbayasgalan/8/orig 2025-09-07T07:28:18.9109417Z * [new branch] gh/tugsbayasgalan/9/base -> origin/gh/tugsbayasgalan/9/base 2025-09-07T07:28:18.9110212Z * [new branch] gh/tugsbayasgalan/9/head -> origin/gh/tugsbayasgalan/9/head 2025-09-07T07:28:18.9111028Z * [new branch] gh/tugsbayasgalan/9/orig -> origin/gh/tugsbayasgalan/9/orig 2025-09-07T07:28:18.9112623Z * [new branch] gh/v0i0/1/base -> origin/gh/v0i0/1/base 2025-09-07T07:28:18.9113475Z * [new branch] gh/v0i0/1/head -> origin/gh/v0i0/1/head 2025-09-07T07:28:18.9114309Z * [new branch] gh/v0i0/1/orig -> origin/gh/v0i0/1/orig 2025-09-07T07:28:18.9115539Z * [new branch] gh/v0i0/4/base -> origin/gh/v0i0/4/base 2025-09-07T07:28:18.9116313Z * [new branch] gh/v0i0/4/head -> origin/gh/v0i0/4/head 2025-09-07T07:28:18.9117227Z * [new branch] gh/v0i0/4/orig -> origin/gh/v0i0/4/orig 2025-09-07T07:28:18.9118433Z * [new branch] gh/v0i0/6/base -> origin/gh/v0i0/6/base 2025-09-07T07:28:18.9119279Z * [new branch] gh/v0i0/6/head -> origin/gh/v0i0/6/head 2025-09-07T07:28:18.9120154Z * [new branch] gh/v0i0/6/orig -> origin/gh/v0i0/6/orig 2025-09-07T07:28:18.9121360Z * [new branch] gh/v0i0/7/base -> origin/gh/v0i0/7/base 2025-09-07T07:28:18.9122272Z * [new branch] gh/v0i0/7/head -> origin/gh/v0i0/7/head 2025-09-07T07:28:18.9123079Z * [new branch] gh/v0i0/7/orig -> origin/gh/v0i0/7/orig 2025-09-07T07:28:18.9124212Z * [new branch] gh/v0i0/8/base -> origin/gh/v0i0/8/base 2025-09-07T07:28:18.9124997Z * [new branch] gh/v0i0/8/head -> origin/gh/v0i0/8/head 2025-09-07T07:28:18.9125761Z * [new branch] gh/v0i0/8/orig -> origin/gh/v0i0/8/orig 2025-09-07T07:28:18.9127025Z * [new branch] gh/v0i0/9/base -> origin/gh/v0i0/9/base 2025-09-07T07:28:18.9127919Z * [new branch] gh/v0i0/9/head -> origin/gh/v0i0/9/head 2025-09-07T07:28:18.9128844Z * [new branch] gh/v0i0/9/orig -> origin/gh/v0i0/9/orig 2025-09-07T07:28:18.9130274Z * [new branch] gh/vkuzo/1/next -> origin/gh/vkuzo/1/next 2025-09-07T07:28:18.9131611Z * [new branch] gh/vkuzo/2/next -> origin/gh/vkuzo/2/next 2025-09-07T07:28:18.9132778Z * [new branch] gh/vkuzo/3/next -> origin/gh/vkuzo/3/next 2025-09-07T07:28:18.9133997Z * [new branch] gh/vkuzo/4/base -> origin/gh/vkuzo/4/base 2025-09-07T07:28:18.9135213Z * [new branch] gh/vkuzo/4/head -> origin/gh/vkuzo/4/head 2025-09-07T07:28:18.9135972Z * [new branch] gh/vkuzo/4/orig -> origin/gh/vkuzo/4/orig 2025-09-07T07:28:18.9137310Z * [new branch] gh/vkuzo/5/base -> origin/gh/vkuzo/5/base 2025-09-07T07:28:18.9138238Z * [new branch] gh/vkuzo/5/head -> origin/gh/vkuzo/5/head 2025-09-07T07:28:18.9139143Z * [new branch] gh/vkuzo/5/orig -> origin/gh/vkuzo/5/orig 2025-09-07T07:28:18.9140459Z * [new branch] gh/vkuzo/6/base -> origin/gh/vkuzo/6/base 2025-09-07T07:28:18.9141288Z * [new branch] gh/vkuzo/6/head -> origin/gh/vkuzo/6/head 2025-09-07T07:28:18.9142138Z * [new branch] gh/vkuzo/6/orig -> origin/gh/vkuzo/6/orig 2025-09-07T07:28:18.9143399Z * [new branch] gh/vkuzo/7/base -> origin/gh/vkuzo/7/base 2025-09-07T07:28:18.9144333Z * [new branch] gh/vkuzo/7/head -> origin/gh/vkuzo/7/head 2025-09-07T07:28:18.9145217Z * [new branch] gh/vkuzo/7/orig -> origin/gh/vkuzo/7/orig 2025-09-07T07:28:18.9146805Z * [new branch] gh/wconstab/419/base -> origin/gh/wconstab/419/base 2025-09-07T07:28:18.9147647Z * [new branch] gh/wconstab/419/head -> origin/gh/wconstab/419/head 2025-09-07T07:28:18.9148510Z * [new branch] gh/wconstab/419/orig -> origin/gh/wconstab/419/orig 2025-09-07T07:28:18.9149758Z * [new branch] gh/wconstab/424/base -> origin/gh/wconstab/424/base 2025-09-07T07:28:18.9150557Z * [new branch] gh/wconstab/424/head -> origin/gh/wconstab/424/head 2025-09-07T07:28:18.9151452Z * [new branch] gh/wconstab/424/orig -> origin/gh/wconstab/424/orig 2025-09-07T07:28:18.9152670Z * [new branch] gh/wconstab/435/base -> origin/gh/wconstab/435/base 2025-09-07T07:28:18.9153561Z * [new branch] gh/wconstab/435/head -> origin/gh/wconstab/435/head 2025-09-07T07:28:18.9154389Z * [new branch] gh/wconstab/435/orig -> origin/gh/wconstab/435/orig 2025-09-07T07:28:18.9155809Z * [new branch] gh/wconstab/438/base -> origin/gh/wconstab/438/base 2025-09-07T07:28:18.9156671Z * [new branch] gh/wconstab/438/head -> origin/gh/wconstab/438/head 2025-09-07T07:28:18.9157485Z * [new branch] gh/wconstab/438/orig -> origin/gh/wconstab/438/orig 2025-09-07T07:28:18.9158710Z * [new branch] gh/wconstab/440/base -> origin/gh/wconstab/440/base 2025-09-07T07:28:18.9159696Z * [new branch] gh/wconstab/440/head -> origin/gh/wconstab/440/head 2025-09-07T07:28:18.9160633Z * [new branch] gh/wconstab/440/orig -> origin/gh/wconstab/440/orig 2025-09-07T07:28:18.9162002Z * [new branch] gh/wconstab/441/base -> origin/gh/wconstab/441/base 2025-09-07T07:28:18.9162797Z * [new branch] gh/wconstab/441/head -> origin/gh/wconstab/441/head 2025-09-07T07:28:18.9163772Z * [new branch] gh/wconstab/441/orig -> origin/gh/wconstab/441/orig 2025-09-07T07:28:18.9165090Z * [new branch] gh/wconstab/442/base -> origin/gh/wconstab/442/base 2025-09-07T07:28:18.9166146Z * [new branch] gh/wconstab/442/head -> origin/gh/wconstab/442/head 2025-09-07T07:28:18.9167029Z * [new branch] gh/wconstab/442/orig -> origin/gh/wconstab/442/orig 2025-09-07T07:28:18.9168211Z * [new branch] gh/wconstab/443/base -> origin/gh/wconstab/443/base 2025-09-07T07:28:18.9169024Z * [new branch] gh/wconstab/443/head -> origin/gh/wconstab/443/head 2025-09-07T07:28:18.9169849Z * [new branch] gh/wconstab/443/orig -> origin/gh/wconstab/443/orig 2025-09-07T07:28:18.9171196Z * [new branch] gh/wconstab/444/base -> origin/gh/wconstab/444/base 2025-09-07T07:28:18.9171978Z * [new branch] gh/wconstab/444/head -> origin/gh/wconstab/444/head 2025-09-07T07:28:18.9172814Z * [new branch] gh/wconstab/444/orig -> origin/gh/wconstab/444/orig 2025-09-07T07:28:18.9174008Z * [new branch] gh/wconstab/445/base -> origin/gh/wconstab/445/base 2025-09-07T07:28:18.9174844Z * [new branch] gh/wconstab/445/head -> origin/gh/wconstab/445/head 2025-09-07T07:28:18.9175773Z * [new branch] gh/wconstab/445/orig -> origin/gh/wconstab/445/orig 2025-09-07T07:28:18.9177403Z * [new branch] gh/wconstab/446/base -> origin/gh/wconstab/446/base 2025-09-07T07:28:18.9178416Z * [new branch] gh/wconstab/446/head -> origin/gh/wconstab/446/head 2025-09-07T07:28:18.9179628Z * [new branch] gh/wconstab/446/orig -> origin/gh/wconstab/446/orig 2025-09-07T07:28:18.9180785Z * [new branch] gh/wconstab/447/base -> origin/gh/wconstab/447/base 2025-09-07T07:28:18.9181765Z * [new branch] gh/wconstab/447/head -> origin/gh/wconstab/447/head 2025-09-07T07:28:18.9182639Z * [new branch] gh/wconstab/447/orig -> origin/gh/wconstab/447/orig 2025-09-07T07:28:18.9184148Z * [new branch] gh/weifengpy/27/base -> origin/gh/weifengpy/27/base 2025-09-07T07:28:18.9185029Z * [new branch] gh/weifengpy/27/head -> origin/gh/weifengpy/27/head 2025-09-07T07:28:18.9185867Z * [new branch] gh/weifengpy/27/orig -> origin/gh/weifengpy/27/orig 2025-09-07T07:28:18.9187041Z * [new branch] gh/weifengpy/30/base -> origin/gh/weifengpy/30/base 2025-09-07T07:28:18.9187959Z * [new branch] gh/weifengpy/30/head -> origin/gh/weifengpy/30/head 2025-09-07T07:28:18.9188776Z * [new branch] gh/weifengpy/30/orig -> origin/gh/weifengpy/30/orig 2025-09-07T07:28:18.9190316Z * [new branch] gh/williamwen42/196/base -> origin/gh/williamwen42/196/base 2025-09-07T07:28:18.9191217Z * [new branch] gh/williamwen42/196/head -> origin/gh/williamwen42/196/head 2025-09-07T07:28:18.9192236Z * [new branch] gh/williamwen42/196/orig -> origin/gh/williamwen42/196/orig 2025-09-07T07:28:18.9193472Z * [new branch] gh/williamwen42/250/base -> origin/gh/williamwen42/250/base 2025-09-07T07:28:18.9194442Z * [new branch] gh/williamwen42/250/head -> origin/gh/williamwen42/250/head 2025-09-07T07:28:18.9195229Z * [new branch] gh/williamwen42/250/orig -> origin/gh/williamwen42/250/orig 2025-09-07T07:28:18.9196533Z * [new branch] gh/williamwen42/258/base -> origin/gh/williamwen42/258/base 2025-09-07T07:28:18.9197414Z * [new branch] gh/williamwen42/258/head -> origin/gh/williamwen42/258/head 2025-09-07T07:28:18.9198328Z * [new branch] gh/williamwen42/258/orig -> origin/gh/williamwen42/258/orig 2025-09-07T07:28:18.9199811Z * [new branch] gh/williamwen42/266/base -> origin/gh/williamwen42/266/base 2025-09-07T07:28:18.9200651Z * [new branch] gh/williamwen42/266/head -> origin/gh/williamwen42/266/head 2025-09-07T07:28:18.9201722Z * [new branch] gh/williamwen42/266/orig -> origin/gh/williamwen42/266/orig 2025-09-07T07:28:18.9202971Z * [new branch] gh/williamwen42/267/base -> origin/gh/williamwen42/267/base 2025-09-07T07:28:18.9203933Z * [new branch] gh/williamwen42/267/head -> origin/gh/williamwen42/267/head 2025-09-07T07:28:18.9204811Z * [new branch] gh/williamwen42/267/orig -> origin/gh/williamwen42/267/orig 2025-09-07T07:28:18.9206080Z * [new branch] gh/williamwen42/270/base -> origin/gh/williamwen42/270/base 2025-09-07T07:28:18.9206980Z * [new branch] gh/williamwen42/270/head -> origin/gh/williamwen42/270/head 2025-09-07T07:28:18.9207905Z * [new branch] gh/williamwen42/270/orig -> origin/gh/williamwen42/270/orig 2025-09-07T07:28:18.9209017Z * [new branch] gh/williamwen42/271/base -> origin/gh/williamwen42/271/base 2025-09-07T07:28:18.9209931Z * [new branch] gh/williamwen42/271/head -> origin/gh/williamwen42/271/head 2025-09-07T07:28:18.9210927Z * [new branch] gh/williamwen42/271/orig -> origin/gh/williamwen42/271/orig 2025-09-07T07:28:18.9212146Z * [new branch] gh/williamwen42/272/base -> origin/gh/williamwen42/272/base 2025-09-07T07:28:18.9213023Z * [new branch] gh/williamwen42/272/head -> origin/gh/williamwen42/272/head 2025-09-07T07:28:18.9213913Z * [new branch] gh/williamwen42/272/orig -> origin/gh/williamwen42/272/orig 2025-09-07T07:28:18.9215079Z * [new branch] gh/williamwen42/274/base -> origin/gh/williamwen42/274/base 2025-09-07T07:28:18.9216015Z * [new branch] gh/williamwen42/274/head -> origin/gh/williamwen42/274/head 2025-09-07T07:28:18.9216858Z * [new branch] gh/williamwen42/274/orig -> origin/gh/williamwen42/274/orig 2025-09-07T07:28:18.9218086Z * [new branch] gh/williamwen42/275/base -> origin/gh/williamwen42/275/base 2025-09-07T07:28:18.9220196Z * [new branch] gh/williamwen42/275/head -> origin/gh/williamwen42/275/head 2025-09-07T07:28:18.9220404Z * [new branch] gh/williamwen42/276/base -> origin/gh/williamwen42/276/base 2025-09-07T07:28:18.9221133Z * [new branch] gh/williamwen42/276/head -> origin/gh/williamwen42/276/head 2025-09-07T07:28:18.9221827Z * [new branch] gh/williamwen42/276/orig -> origin/gh/williamwen42/276/orig 2025-09-07T07:28:18.9223058Z * [new branch] gh/williamwen42/277/base -> origin/gh/williamwen42/277/base 2025-09-07T07:28:18.9224002Z * [new branch] gh/williamwen42/277/head -> origin/gh/williamwen42/277/head 2025-09-07T07:28:18.9224829Z * [new branch] gh/williamwen42/277/orig -> origin/gh/williamwen42/277/orig 2025-09-07T07:28:18.9226015Z * [new branch] gh/williamwen42/278/base -> origin/gh/williamwen42/278/base 2025-09-07T07:28:18.9226896Z * [new branch] gh/williamwen42/278/head -> origin/gh/williamwen42/278/head 2025-09-07T07:28:18.9227727Z * [new branch] gh/williamwen42/278/orig -> origin/gh/williamwen42/278/orig 2025-09-07T07:28:18.9228960Z * [new branch] gh/williamwen42/279/base -> origin/gh/williamwen42/279/base 2025-09-07T07:28:18.9229804Z * [new branch] gh/williamwen42/279/head -> origin/gh/williamwen42/279/head 2025-09-07T07:28:18.9230682Z * [new branch] gh/williamwen42/279/orig -> origin/gh/williamwen42/279/orig 2025-09-07T07:28:18.9231912Z * [new branch] gh/williamwen42/280/base -> origin/gh/williamwen42/280/base 2025-09-07T07:28:18.9232761Z * [new branch] gh/williamwen42/280/head -> origin/gh/williamwen42/280/head 2025-09-07T07:28:18.9233619Z * [new branch] gh/williamwen42/280/orig -> origin/gh/williamwen42/280/orig 2025-09-07T07:28:18.9234852Z * [new branch] gh/williamwen42/281/base -> origin/gh/williamwen42/281/base 2025-09-07T07:28:18.9235667Z * [new branch] gh/williamwen42/281/head -> origin/gh/williamwen42/281/head 2025-09-07T07:28:18.9236508Z * [new branch] gh/williamwen42/281/orig -> origin/gh/williamwen42/281/orig 2025-09-07T07:28:18.9237919Z * [new branch] gh/williamwen42/282/base -> origin/gh/williamwen42/282/base 2025-09-07T07:28:18.9238926Z * [new branch] gh/williamwen42/282/head -> origin/gh/williamwen42/282/head 2025-09-07T07:28:18.9239804Z * [new branch] gh/williamwen42/282/orig -> origin/gh/williamwen42/282/orig 2025-09-07T07:28:18.9241134Z * [new branch] gh/williamwen42/283/base -> origin/gh/williamwen42/283/base 2025-09-07T07:28:18.9242150Z * [new branch] gh/williamwen42/283/head -> origin/gh/williamwen42/283/head 2025-09-07T07:28:18.9242916Z * [new branch] gh/williamwen42/283/orig -> origin/gh/williamwen42/283/orig 2025-09-07T07:28:18.9244384Z * [new branch] gh/williamwen42/284/base -> origin/gh/williamwen42/284/base 2025-09-07T07:28:18.9245245Z * [new branch] gh/williamwen42/284/head -> origin/gh/williamwen42/284/head 2025-09-07T07:28:18.9246180Z * [new branch] gh/williamwen42/284/orig -> origin/gh/williamwen42/284/orig 2025-09-07T07:28:18.9247241Z * [new branch] gh/williamwen42/285/base -> origin/gh/williamwen42/285/base 2025-09-07T07:28:18.9248194Z * [new branch] gh/williamwen42/285/head -> origin/gh/williamwen42/285/head 2025-09-07T07:28:18.9249026Z * [new branch] gh/williamwen42/285/orig -> origin/gh/williamwen42/285/orig 2025-09-07T07:28:18.9250173Z * [new branch] gh/williamwen42/286/base -> origin/gh/williamwen42/286/base 2025-09-07T07:28:18.9250983Z * [new branch] gh/williamwen42/286/head -> origin/gh/williamwen42/286/head 2025-09-07T07:28:18.9251853Z * [new branch] gh/williamwen42/286/orig -> origin/gh/williamwen42/286/orig 2025-09-07T07:28:18.9253104Z * [new branch] gh/williamwen42/287/base -> origin/gh/williamwen42/287/base 2025-09-07T07:28:18.9253994Z * [new branch] gh/williamwen42/287/head -> origin/gh/williamwen42/287/head 2025-09-07T07:28:18.9254892Z * [new branch] gh/williamwen42/287/orig -> origin/gh/williamwen42/287/orig 2025-09-07T07:28:18.9256190Z * [new branch] gh/williamwen42/288/base -> origin/gh/williamwen42/288/base 2025-09-07T07:28:18.9257031Z * [new branch] gh/williamwen42/288/head -> origin/gh/williamwen42/288/head 2025-09-07T07:28:18.9257906Z * [new branch] gh/williamwen42/288/orig -> origin/gh/williamwen42/288/orig 2025-09-07T07:28:18.9259114Z * [new branch] gh/williamwen42/289/base -> origin/gh/williamwen42/289/base 2025-09-07T07:28:18.9260016Z * [new branch] gh/williamwen42/289/head -> origin/gh/williamwen42/289/head 2025-09-07T07:28:18.9260854Z * [new branch] gh/williamwen42/289/orig -> origin/gh/williamwen42/289/orig 2025-09-07T07:28:18.9262607Z * [new branch] gh/wychi/1/base -> origin/gh/wychi/1/base 2025-09-07T07:28:18.9263543Z * [new branch] gh/wychi/1/head -> origin/gh/wychi/1/head 2025-09-07T07:28:18.9264478Z * [new branch] gh/wychi/1/orig -> origin/gh/wychi/1/orig 2025-09-07T07:28:18.9269324Z * [new branch] gh/xmfan/169/base -> origin/gh/xmfan/169/base 2025-09-07T07:28:18.9270328Z * [new branch] gh/xmfan/169/head -> origin/gh/xmfan/169/head 2025-09-07T07:28:18.9271485Z * [new branch] gh/xmfan/170/base -> origin/gh/xmfan/170/base 2025-09-07T07:28:18.9272256Z * [new branch] gh/xmfan/170/head -> origin/gh/xmfan/170/head 2025-09-07T07:28:18.9273577Z * [new branch] gh/xmfan/18/base -> origin/gh/xmfan/18/base 2025-09-07T07:28:18.9274494Z * [new branch] gh/xmfan/18/head -> origin/gh/xmfan/18/head 2025-09-07T07:28:18.9275661Z * [new branch] gh/xmfan/229/base -> origin/gh/xmfan/229/base 2025-09-07T07:28:18.9276524Z * [new branch] gh/xmfan/229/head -> origin/gh/xmfan/229/head 2025-09-07T07:28:18.9277384Z * [new branch] gh/xmfan/229/orig -> origin/gh/xmfan/229/orig 2025-09-07T07:28:18.9278539Z * [new branch] gh/xmfan/237/base -> origin/gh/xmfan/237/base 2025-09-07T07:28:18.9279414Z * [new branch] gh/xmfan/237/head -> origin/gh/xmfan/237/head 2025-09-07T07:28:18.9280708Z * [new branch] gh/xmfan/237/orig -> origin/gh/xmfan/237/orig 2025-09-07T07:28:18.9281769Z * [new branch] gh/xmfan/244/base -> origin/gh/xmfan/244/base 2025-09-07T07:28:18.9282614Z * [new branch] gh/xmfan/244/head -> origin/gh/xmfan/244/head 2025-09-07T07:28:18.9283451Z * [new branch] gh/xmfan/244/orig -> origin/gh/xmfan/244/orig 2025-09-07T07:28:18.9284600Z * [new branch] gh/xmfan/246/base -> origin/gh/xmfan/246/base 2025-09-07T07:28:18.9285456Z * [new branch] gh/xmfan/246/head -> origin/gh/xmfan/246/head 2025-09-07T07:28:18.9286314Z * [new branch] gh/xmfan/246/orig -> origin/gh/xmfan/246/orig 2025-09-07T07:28:18.9287538Z * [new branch] gh/xmfan/253/base -> origin/gh/xmfan/253/base 2025-09-07T07:28:18.9288372Z * [new branch] gh/xmfan/253/head -> origin/gh/xmfan/253/head 2025-09-07T07:28:18.9289237Z * [new branch] gh/xmfan/253/orig -> origin/gh/xmfan/253/orig 2025-09-07T07:28:18.9290527Z * [new branch] gh/xmfan/254/base -> origin/gh/xmfan/254/base 2025-09-07T07:28:18.9291403Z * [new branch] gh/xmfan/254/head -> origin/gh/xmfan/254/head 2025-09-07T07:28:18.9292220Z * [new branch] gh/xmfan/254/orig -> origin/gh/xmfan/254/orig 2025-09-07T07:28:18.9293448Z * [new branch] gh/xmfan/260/base -> origin/gh/xmfan/260/base 2025-09-07T07:28:18.9294305Z * [new branch] gh/xmfan/260/head -> origin/gh/xmfan/260/head 2025-09-07T07:28:18.9295129Z * [new branch] gh/xmfan/260/orig -> origin/gh/xmfan/260/orig 2025-09-07T07:28:18.9296276Z * [new branch] gh/xmfan/262/base -> origin/gh/xmfan/262/base 2025-09-07T07:28:18.9297112Z * [new branch] gh/xmfan/262/head -> origin/gh/xmfan/262/head 2025-09-07T07:28:18.9298017Z * [new branch] gh/xmfan/262/orig -> origin/gh/xmfan/262/orig 2025-09-07T07:28:18.9299537Z * [new branch] gh/xmfan/263/base -> origin/gh/xmfan/263/base 2025-09-07T07:28:18.9300367Z * [new branch] gh/xmfan/263/head -> origin/gh/xmfan/263/head 2025-09-07T07:28:18.9301162Z * [new branch] gh/xmfan/263/orig -> origin/gh/xmfan/263/orig 2025-09-07T07:28:18.9302474Z * [new branch] gh/xmfan/264/base -> origin/gh/xmfan/264/base 2025-09-07T07:28:18.9303301Z * [new branch] gh/xmfan/264/head -> origin/gh/xmfan/264/head 2025-09-07T07:28:18.9304187Z * [new branch] gh/xmfan/264/orig -> origin/gh/xmfan/264/orig 2025-09-07T07:28:18.9305340Z * [new branch] gh/xmfan/274/base -> origin/gh/xmfan/274/base 2025-09-07T07:28:18.9306153Z * [new branch] gh/xmfan/274/head -> origin/gh/xmfan/274/head 2025-09-07T07:28:18.9307049Z * [new branch] gh/xmfan/274/orig -> origin/gh/xmfan/274/orig 2025-09-07T07:28:18.9308322Z * [new branch] gh/xmfan/276/base -> origin/gh/xmfan/276/base 2025-09-07T07:28:18.9309139Z * [new branch] gh/xmfan/276/head -> origin/gh/xmfan/276/head 2025-09-07T07:28:18.9310048Z * [new branch] gh/xmfan/276/orig -> origin/gh/xmfan/276/orig 2025-09-07T07:28:18.9311229Z * [new branch] gh/xmfan/277/base -> origin/gh/xmfan/277/base 2025-09-07T07:28:18.9312025Z * [new branch] gh/xmfan/277/head -> origin/gh/xmfan/277/head 2025-09-07T07:28:18.9312870Z * [new branch] gh/xmfan/277/orig -> origin/gh/xmfan/277/orig 2025-09-07T07:28:18.9313981Z * [new branch] gh/xmfan/278/base -> origin/gh/xmfan/278/base 2025-09-07T07:28:18.9314819Z * [new branch] gh/xmfan/278/head -> origin/gh/xmfan/278/head 2025-09-07T07:28:18.9315806Z * [new branch] gh/xmfan/278/orig -> origin/gh/xmfan/278/orig 2025-09-07T07:28:18.9317274Z * [new branch] gh/xmfan/279/base -> origin/gh/xmfan/279/base 2025-09-07T07:28:18.9318053Z * [new branch] gh/xmfan/279/head -> origin/gh/xmfan/279/head 2025-09-07T07:28:18.9318870Z * [new branch] gh/xmfan/279/orig -> origin/gh/xmfan/279/orig 2025-09-07T07:28:18.9320163Z * [new branch] gh/xmfan/280/base -> origin/gh/xmfan/280/base 2025-09-07T07:28:18.9321013Z * [new branch] gh/xmfan/280/head -> origin/gh/xmfan/280/head 2025-09-07T07:28:18.9321881Z * [new branch] gh/xmfan/280/orig -> origin/gh/xmfan/280/orig 2025-09-07T07:28:18.9323108Z * [new branch] gh/xmfan/281/base -> origin/gh/xmfan/281/base 2025-09-07T07:28:18.9323944Z * [new branch] gh/xmfan/281/head -> origin/gh/xmfan/281/head 2025-09-07T07:28:18.9324769Z * [new branch] gh/xmfan/281/orig -> origin/gh/xmfan/281/orig 2025-09-07T07:28:18.9325981Z * [new branch] gh/xmfan/282/base -> origin/gh/xmfan/282/base 2025-09-07T07:28:18.9326822Z * [new branch] gh/xmfan/282/head -> origin/gh/xmfan/282/head 2025-09-07T07:28:18.9328048Z * [new branch] gh/xmfan/283/base -> origin/gh/xmfan/283/base 2025-09-07T07:28:18.9328908Z * [new branch] gh/xmfan/283/head -> origin/gh/xmfan/283/head 2025-09-07T07:28:18.9329800Z * [new branch] gh/xmfan/283/orig -> origin/gh/xmfan/283/orig 2025-09-07T07:28:18.9331275Z * [new branch] gh/xuanzhang816/14/base -> origin/gh/xuanzhang816/14/base 2025-09-07T07:28:18.9334836Z * [new branch] gh/xuanzhang816/14/head -> origin/gh/xuanzhang816/14/head 2025-09-07T07:28:18.9335683Z * [new branch] gh/xuanzhang816/14/orig -> origin/gh/xuanzhang816/14/orig 2025-09-07T07:28:18.9336894Z * [new branch] gh/xuanzhang816/19/base -> origin/gh/xuanzhang816/19/base 2025-09-07T07:28:18.9337753Z * [new branch] gh/xuanzhang816/19/head -> origin/gh/xuanzhang816/19/head 2025-09-07T07:28:18.9338650Z * [new branch] gh/xuanzhang816/19/orig -> origin/gh/xuanzhang816/19/orig 2025-09-07T07:28:18.9339837Z * [new branch] gh/xuanzhang816/22/base -> origin/gh/xuanzhang816/22/base 2025-09-07T07:28:18.9340634Z * [new branch] gh/xuanzhang816/22/head -> origin/gh/xuanzhang816/22/head 2025-09-07T07:28:18.9341502Z * [new branch] gh/xuanzhang816/22/orig -> origin/gh/xuanzhang816/22/orig 2025-09-07T07:28:18.9342633Z * [new branch] gh/xuanzhang816/23/base -> origin/gh/xuanzhang816/23/base 2025-09-07T07:28:18.9343488Z * [new branch] gh/xuanzhang816/23/head -> origin/gh/xuanzhang816/23/head 2025-09-07T07:28:18.9344354Z * [new branch] gh/xuanzhang816/23/orig -> origin/gh/xuanzhang816/23/orig 2025-09-07T07:28:18.9345548Z * [new branch] gh/xuanzhang816/24/base -> origin/gh/xuanzhang816/24/base 2025-09-07T07:28:18.9346373Z * [new branch] gh/xuanzhang816/24/head -> origin/gh/xuanzhang816/24/head 2025-09-07T07:28:18.9347226Z * [new branch] gh/xuanzhang816/24/orig -> origin/gh/xuanzhang816/24/orig 2025-09-07T07:28:18.9348402Z * [new branch] gh/xuanzhang816/25/base -> origin/gh/xuanzhang816/25/base 2025-09-07T07:28:18.9349221Z * [new branch] gh/xuanzhang816/25/head -> origin/gh/xuanzhang816/25/head 2025-09-07T07:28:18.9350064Z * [new branch] gh/xuanzhang816/25/orig -> origin/gh/xuanzhang816/25/orig 2025-09-07T07:28:18.9351211Z * [new branch] gh/xuanzhang816/26/base -> origin/gh/xuanzhang816/26/base 2025-09-07T07:28:18.9351997Z * [new branch] gh/xuanzhang816/26/head -> origin/gh/xuanzhang816/26/head 2025-09-07T07:28:18.9352983Z * [new branch] gh/xuanzhang816/26/orig -> origin/gh/xuanzhang816/26/orig 2025-09-07T07:28:18.9354445Z * [new branch] gh/yanbing-j/11/base -> origin/gh/yanbing-j/11/base 2025-09-07T07:28:18.9355315Z * [new branch] gh/yanbing-j/11/head -> origin/gh/yanbing-j/11/head 2025-09-07T07:28:18.9356438Z * [new branch] gh/yanbing-j/11/orig -> origin/gh/yanbing-j/11/orig 2025-09-07T07:28:18.9357647Z * [new branch] gh/yanbing-j/12/base -> origin/gh/yanbing-j/12/base 2025-09-07T07:28:18.9358540Z * [new branch] gh/yanbing-j/12/head -> origin/gh/yanbing-j/12/head 2025-09-07T07:28:18.9359432Z * [new branch] gh/yanbing-j/12/orig -> origin/gh/yanbing-j/12/orig 2025-09-07T07:28:18.9360611Z * [new branch] gh/yanbing-j/13/base -> origin/gh/yanbing-j/13/base 2025-09-07T07:28:18.9361544Z * [new branch] gh/yanbing-j/13/head -> origin/gh/yanbing-j/13/head 2025-09-07T07:28:18.9362337Z * [new branch] gh/yanbing-j/13/orig -> origin/gh/yanbing-j/13/orig 2025-09-07T07:28:18.9363501Z * [new branch] gh/yanbing-j/14/base -> origin/gh/yanbing-j/14/base 2025-09-07T07:28:18.9364313Z * [new branch] gh/yanbing-j/14/head -> origin/gh/yanbing-j/14/head 2025-09-07T07:28:18.9365248Z * [new branch] gh/yanbing-j/14/orig -> origin/gh/yanbing-j/14/orig 2025-09-07T07:28:18.9366376Z * [new branch] gh/yanbing-j/15/base -> origin/gh/yanbing-j/15/base 2025-09-07T07:28:18.9367288Z * [new branch] gh/yanbing-j/15/head -> origin/gh/yanbing-j/15/head 2025-09-07T07:28:18.9368111Z * [new branch] gh/yanbing-j/15/orig -> origin/gh/yanbing-j/15/orig 2025-09-07T07:28:18.9369207Z * [new branch] gh/yanbing-j/18/base -> origin/gh/yanbing-j/18/base 2025-09-07T07:28:18.9370315Z * [new branch] gh/yanbing-j/18/head -> origin/gh/yanbing-j/18/head 2025-09-07T07:28:18.9371182Z * [new branch] gh/yanbing-j/18/orig -> origin/gh/yanbing-j/18/orig 2025-09-07T07:28:18.9372291Z * [new branch] gh/yanbing-j/19/base -> origin/gh/yanbing-j/19/base 2025-09-07T07:28:18.9373185Z * [new branch] gh/yanbing-j/19/head -> origin/gh/yanbing-j/19/head 2025-09-07T07:28:18.9373993Z * [new branch] gh/yanbing-j/19/orig -> origin/gh/yanbing-j/19/orig 2025-09-07T07:28:18.9375230Z * [new branch] gh/yanbing-j/20/base -> origin/gh/yanbing-j/20/base 2025-09-07T07:28:18.9376083Z * [new branch] gh/yanbing-j/20/head -> origin/gh/yanbing-j/20/head 2025-09-07T07:28:18.9376920Z * [new branch] gh/yanbing-j/20/orig -> origin/gh/yanbing-j/20/orig 2025-09-07T07:28:18.9378098Z * [new branch] gh/yanbing-j/21/base -> origin/gh/yanbing-j/21/base 2025-09-07T07:28:18.9378956Z * [new branch] gh/yanbing-j/21/head -> origin/gh/yanbing-j/21/head 2025-09-07T07:28:18.9380114Z * [new branch] gh/yanbing-j/22/base -> origin/gh/yanbing-j/22/base 2025-09-07T07:28:18.9381007Z * [new branch] gh/yanbing-j/22/head -> origin/gh/yanbing-j/22/head 2025-09-07T07:28:18.9381897Z * [new branch] gh/yanbing-j/22/orig -> origin/gh/yanbing-j/22/orig 2025-09-07T07:28:18.9383081Z * [new branch] gh/yanbing-j/23/base -> origin/gh/yanbing-j/23/base 2025-09-07T07:28:18.9383891Z * [new branch] gh/yanbing-j/23/head -> origin/gh/yanbing-j/23/head 2025-09-07T07:28:18.9384761Z * [new branch] gh/yanbing-j/23/orig -> origin/gh/yanbing-j/23/orig 2025-09-07T07:28:18.9385913Z * [new branch] gh/yanbing-j/24/base -> origin/gh/yanbing-j/24/base 2025-09-07T07:28:18.9386700Z * [new branch] gh/yanbing-j/24/head -> origin/gh/yanbing-j/24/head 2025-09-07T07:28:18.9387704Z * [new branch] gh/yanbing-j/24/orig -> origin/gh/yanbing-j/24/orig 2025-09-07T07:28:18.9388752Z * [new branch] gh/yanbing-j/25/base -> origin/gh/yanbing-j/25/base 2025-09-07T07:28:18.9389612Z * [new branch] gh/yanbing-j/25/head -> origin/gh/yanbing-j/25/head 2025-09-07T07:28:18.9390467Z * [new branch] gh/yanbing-j/25/orig -> origin/gh/yanbing-j/25/orig 2025-09-07T07:28:18.9391646Z * [new branch] gh/yanbing-j/26/base -> origin/gh/yanbing-j/26/base 2025-09-07T07:28:18.9392519Z * [new branch] gh/yanbing-j/26/head -> origin/gh/yanbing-j/26/head 2025-09-07T07:28:18.9393405Z * [new branch] gh/yanbing-j/26/orig -> origin/gh/yanbing-j/26/orig 2025-09-07T07:28:18.9394523Z * [new branch] gh/yanbing-j/36/base -> origin/gh/yanbing-j/36/base 2025-09-07T07:28:18.9395366Z * [new branch] gh/yanbing-j/36/head -> origin/gh/yanbing-j/36/head 2025-09-07T07:28:18.9396263Z * [new branch] gh/yanbing-j/36/orig -> origin/gh/yanbing-j/36/orig 2025-09-07T07:28:18.9397439Z * [new branch] gh/yanbing-j/37/base -> origin/gh/yanbing-j/37/base 2025-09-07T07:28:18.9398313Z * [new branch] gh/yanbing-j/37/head -> origin/gh/yanbing-j/37/head 2025-09-07T07:28:18.9399189Z * [new branch] gh/yanbing-j/37/orig -> origin/gh/yanbing-j/37/orig 2025-09-07T07:28:18.9400566Z * [new branch] gh/yangw-dev/12/base -> origin/gh/yangw-dev/12/base 2025-09-07T07:28:18.9401467Z * [new branch] gh/yangw-dev/12/head -> origin/gh/yangw-dev/12/head 2025-09-07T07:28:18.9402467Z * [new branch] gh/yangw-dev/12/orig -> origin/gh/yangw-dev/12/orig 2025-09-07T07:28:18.9403709Z * [new branch] gh/yangw-dev/13/base -> origin/gh/yangw-dev/13/base 2025-09-07T07:28:18.9404701Z * [new branch] gh/yangw-dev/13/head -> origin/gh/yangw-dev/13/head 2025-09-07T07:28:18.9405578Z * [new branch] gh/yangw-dev/13/orig -> origin/gh/yangw-dev/13/orig 2025-09-07T07:28:18.9406707Z * [new branch] gh/yangw-dev/14/base -> origin/gh/yangw-dev/14/base 2025-09-07T07:28:18.9416218Z * [new branch] gh/yangw-dev/14/head -> origin/gh/yangw-dev/14/head 2025-09-07T07:28:18.9416717Z * [new branch] gh/yangw-dev/14/orig -> origin/gh/yangw-dev/14/orig 2025-09-07T07:28:18.9417186Z * [new branch] gh/yangw-dev/15/base -> origin/gh/yangw-dev/15/base 2025-09-07T07:28:18.9417652Z * [new branch] gh/yangw-dev/15/head -> origin/gh/yangw-dev/15/head 2025-09-07T07:28:18.9418117Z * [new branch] gh/yangw-dev/15/orig -> origin/gh/yangw-dev/15/orig 2025-09-07T07:28:18.9418760Z * [new branch] gh/yangw-dev/16/base -> origin/gh/yangw-dev/16/base 2025-09-07T07:28:18.9419230Z * [new branch] gh/yangw-dev/16/head -> origin/gh/yangw-dev/16/head 2025-09-07T07:28:18.9419700Z * [new branch] gh/yangw-dev/16/orig -> origin/gh/yangw-dev/16/orig 2025-09-07T07:28:18.9420166Z * [new branch] gh/yangw-dev/17/base -> origin/gh/yangw-dev/17/base 2025-09-07T07:28:18.9420637Z * [new branch] gh/yangw-dev/17/head -> origin/gh/yangw-dev/17/head 2025-09-07T07:28:18.9421100Z * [new branch] gh/yangw-dev/17/orig -> origin/gh/yangw-dev/17/orig 2025-09-07T07:28:18.9421563Z * [new branch] gh/yangw-dev/18/base -> origin/gh/yangw-dev/18/base 2025-09-07T07:28:18.9422022Z * [new branch] gh/yangw-dev/18/head -> origin/gh/yangw-dev/18/head 2025-09-07T07:28:18.9422492Z * [new branch] gh/yangw-dev/18/orig -> origin/gh/yangw-dev/18/orig 2025-09-07T07:28:18.9422960Z * [new branch] gh/yangw-dev/19/base -> origin/gh/yangw-dev/19/base 2025-09-07T07:28:18.9423526Z * [new branch] gh/yangw-dev/19/head -> origin/gh/yangw-dev/19/head 2025-09-07T07:28:18.9424016Z * [new branch] gh/yangw-dev/19/orig -> origin/gh/yangw-dev/19/orig 2025-09-07T07:28:18.9424477Z * [new branch] gh/yangw-dev/20/base -> origin/gh/yangw-dev/20/base 2025-09-07T07:28:18.9424993Z * [new branch] gh/yangw-dev/20/head -> origin/gh/yangw-dev/20/head 2025-09-07T07:28:18.9425878Z * [new branch] gh/yangw-dev/20/orig -> origin/gh/yangw-dev/20/orig 2025-09-07T07:28:18.9427001Z * [new branch] gh/yangw-dev/21/base -> origin/gh/yangw-dev/21/base 2025-09-07T07:28:18.9427922Z * [new branch] gh/yangw-dev/21/head -> origin/gh/yangw-dev/21/head 2025-09-07T07:28:18.9428733Z * [new branch] gh/yangw-dev/21/orig -> origin/gh/yangw-dev/21/orig 2025-09-07T07:28:18.9429904Z * [new branch] gh/yangw-dev/22/base -> origin/gh/yangw-dev/22/base 2025-09-07T07:28:18.9430872Z * [new branch] gh/yangw-dev/22/head -> origin/gh/yangw-dev/22/head 2025-09-07T07:28:18.9431694Z * [new branch] gh/yangw-dev/22/orig -> origin/gh/yangw-dev/22/orig 2025-09-07T07:28:18.9432821Z * [new branch] gh/yangw-dev/23/base -> origin/gh/yangw-dev/23/base 2025-09-07T07:28:18.9433683Z * [new branch] gh/yangw-dev/23/head -> origin/gh/yangw-dev/23/head 2025-09-07T07:28:18.9434773Z * [new branch] gh/yangw-dev/23/orig -> origin/gh/yangw-dev/23/orig 2025-09-07T07:28:18.9435977Z * [new branch] gh/yangw-dev/24/base -> origin/gh/yangw-dev/24/base 2025-09-07T07:28:18.9436842Z * [new branch] gh/yangw-dev/24/head -> origin/gh/yangw-dev/24/head 2025-09-07T07:28:18.9437628Z * [new branch] gh/yangw-dev/24/orig -> origin/gh/yangw-dev/24/orig 2025-09-07T07:28:18.9438868Z * [new branch] gh/yangw-dev/25/base -> origin/gh/yangw-dev/25/base 2025-09-07T07:28:18.9439925Z * [new branch] gh/yangw-dev/25/head -> origin/gh/yangw-dev/25/head 2025-09-07T07:28:18.9440777Z * [new branch] gh/yangw-dev/25/orig -> origin/gh/yangw-dev/25/orig 2025-09-07T07:28:18.9441979Z * [new branch] gh/yangw-dev/26/base -> origin/gh/yangw-dev/26/base 2025-09-07T07:28:18.9442806Z * [new branch] gh/yangw-dev/26/head -> origin/gh/yangw-dev/26/head 2025-09-07T07:28:18.9443616Z * [new branch] gh/yangw-dev/26/orig -> origin/gh/yangw-dev/26/orig 2025-09-07T07:28:18.9444839Z * [new branch] gh/yangw-dev/27/base -> origin/gh/yangw-dev/27/base 2025-09-07T07:28:18.9445678Z * [new branch] gh/yangw-dev/27/head -> origin/gh/yangw-dev/27/head 2025-09-07T07:28:18.9446541Z * [new branch] gh/yangw-dev/27/orig -> origin/gh/yangw-dev/27/orig 2025-09-07T07:28:18.9447966Z * [new branch] gh/ydwu4/233/base -> origin/gh/ydwu4/233/base 2025-09-07T07:28:18.9448879Z * [new branch] gh/ydwu4/233/head -> origin/gh/ydwu4/233/head 2025-09-07T07:28:18.9449699Z * [new branch] gh/ydwu4/233/orig -> origin/gh/ydwu4/233/orig 2025-09-07T07:28:18.9451159Z * [new branch] gh/ydwu4/246/base -> origin/gh/ydwu4/246/base 2025-09-07T07:28:18.9452133Z * [new branch] gh/ydwu4/246/head -> origin/gh/ydwu4/246/head 2025-09-07T07:28:18.9452937Z * [new branch] gh/ydwu4/246/orig -> origin/gh/ydwu4/246/orig 2025-09-07T07:28:18.9454213Z * [new branch] gh/ydwu4/253/base -> origin/gh/ydwu4/253/base 2025-09-07T07:28:18.9455145Z * [new branch] gh/ydwu4/253/head -> origin/gh/ydwu4/253/head 2025-09-07T07:28:18.9456051Z * [new branch] gh/ydwu4/253/orig -> origin/gh/ydwu4/253/orig 2025-09-07T07:28:18.9457212Z * [new branch] gh/ydwu4/255/base -> origin/gh/ydwu4/255/base 2025-09-07T07:28:18.9458139Z * [new branch] gh/ydwu4/255/head -> origin/gh/ydwu4/255/head 2025-09-07T07:28:18.9458886Z * [new branch] gh/ydwu4/255/orig -> origin/gh/ydwu4/255/orig 2025-09-07T07:28:18.9460175Z * [new branch] gh/ydwu4/259/base -> origin/gh/ydwu4/259/base 2025-09-07T07:28:18.9460986Z * [new branch] gh/ydwu4/259/head -> origin/gh/ydwu4/259/head 2025-09-07T07:28:18.9461841Z * [new branch] gh/ydwu4/259/orig -> origin/gh/ydwu4/259/orig 2025-09-07T07:28:18.9463176Z * [new branch] gh/ydwu4/262/base -> origin/gh/ydwu4/262/base 2025-09-07T07:28:18.9464057Z * [new branch] gh/ydwu4/262/head -> origin/gh/ydwu4/262/head 2025-09-07T07:28:18.9464896Z * [new branch] gh/ydwu4/262/orig -> origin/gh/ydwu4/262/orig 2025-09-07T07:28:18.9466348Z * [new branch] gh/ydwu4/263/base -> origin/gh/ydwu4/263/base 2025-09-07T07:28:18.9467230Z * [new branch] gh/ydwu4/263/head -> origin/gh/ydwu4/263/head 2025-09-07T07:28:18.9468359Z * [new branch] gh/ydwu4/263/orig -> origin/gh/ydwu4/263/orig 2025-09-07T07:28:18.9469712Z * [new branch] gh/ydwu4/269/base -> origin/gh/ydwu4/269/base 2025-09-07T07:28:18.9470558Z * [new branch] gh/ydwu4/269/head -> origin/gh/ydwu4/269/head 2025-09-07T07:28:18.9471319Z * [new branch] gh/ydwu4/269/orig -> origin/gh/ydwu4/269/orig 2025-09-07T07:28:18.9472578Z * [new branch] gh/ydwu4/270/base -> origin/gh/ydwu4/270/base 2025-09-07T07:28:18.9473458Z * [new branch] gh/ydwu4/270/head -> origin/gh/ydwu4/270/head 2025-09-07T07:28:18.9474379Z * [new branch] gh/ydwu4/270/orig -> origin/gh/ydwu4/270/orig 2025-09-07T07:28:18.9475821Z * [new branch] gh/ydwu4/272/base -> origin/gh/ydwu4/272/base 2025-09-07T07:28:18.9476778Z * [new branch] gh/ydwu4/272/head -> origin/gh/ydwu4/272/head 2025-09-07T07:28:18.9477643Z * [new branch] gh/ydwu4/272/orig -> origin/gh/ydwu4/272/orig 2025-09-07T07:28:18.9478706Z * [new branch] gh/ydwu4/275/base -> origin/gh/ydwu4/275/base 2025-09-07T07:28:18.9479728Z * [new branch] gh/ydwu4/275/head -> origin/gh/ydwu4/275/head 2025-09-07T07:28:18.9480628Z * [new branch] gh/ydwu4/275/orig -> origin/gh/ydwu4/275/orig 2025-09-07T07:28:18.9481738Z * [new branch] gh/ydwu4/276/base -> origin/gh/ydwu4/276/base 2025-09-07T07:28:18.9482628Z * [new branch] gh/ydwu4/276/head -> origin/gh/ydwu4/276/head 2025-09-07T07:28:18.9483413Z * [new branch] gh/ydwu4/276/orig -> origin/gh/ydwu4/276/orig 2025-09-07T07:28:18.9484741Z * [new branch] gh/ydwu4/279/base -> origin/gh/ydwu4/279/base 2025-09-07T07:28:18.9485743Z * [new branch] gh/ydwu4/279/head -> origin/gh/ydwu4/279/head 2025-09-07T07:28:18.9486637Z * [new branch] gh/ydwu4/279/orig -> origin/gh/ydwu4/279/orig 2025-09-07T07:28:18.9488160Z * [new branch] gh/ydwu4/283/base -> origin/gh/ydwu4/283/base 2025-09-07T07:28:18.9488992Z * [new branch] gh/ydwu4/283/head -> origin/gh/ydwu4/283/head 2025-09-07T07:28:18.9489888Z * [new branch] gh/ydwu4/283/orig -> origin/gh/ydwu4/283/orig 2025-09-07T07:28:18.9491063Z * [new branch] gh/ydwu4/289/base -> origin/gh/ydwu4/289/base 2025-09-07T07:28:18.9491955Z * [new branch] gh/ydwu4/289/head -> origin/gh/ydwu4/289/head 2025-09-07T07:28:18.9492773Z * [new branch] gh/ydwu4/289/orig -> origin/gh/ydwu4/289/orig 2025-09-07T07:28:18.9494206Z * [new branch] gh/ydwu4/290/base -> origin/gh/ydwu4/290/base 2025-09-07T07:28:18.9494987Z * [new branch] gh/ydwu4/290/head -> origin/gh/ydwu4/290/head 2025-09-07T07:28:18.9495780Z * [new branch] gh/ydwu4/290/orig -> origin/gh/ydwu4/290/orig 2025-09-07T07:28:18.9497378Z * [new branch] gh/ydwu4/291/base -> origin/gh/ydwu4/291/base 2025-09-07T07:28:18.9498275Z * [new branch] gh/ydwu4/291/head -> origin/gh/ydwu4/291/head 2025-09-07T07:28:18.9499178Z * [new branch] gh/ydwu4/291/orig -> origin/gh/ydwu4/291/orig 2025-09-07T07:28:18.9500544Z * [new branch] gh/ydwu4/292/base -> origin/gh/ydwu4/292/base 2025-09-07T07:28:18.9501324Z * [new branch] gh/ydwu4/292/head -> origin/gh/ydwu4/292/head 2025-09-07T07:28:18.9502114Z * [new branch] gh/ydwu4/292/orig -> origin/gh/ydwu4/292/orig 2025-09-07T07:28:18.9503527Z * [new branch] gh/ydwu4/293/base -> origin/gh/ydwu4/293/base 2025-09-07T07:28:18.9504371Z * [new branch] gh/ydwu4/293/head -> origin/gh/ydwu4/293/head 2025-09-07T07:28:18.9505388Z * [new branch] gh/ydwu4/293/orig -> origin/gh/ydwu4/293/orig 2025-09-07T07:28:18.9506620Z * [new branch] gh/ydwu4/294/base -> origin/gh/ydwu4/294/base 2025-09-07T07:28:18.9507477Z * [new branch] gh/ydwu4/294/head -> origin/gh/ydwu4/294/head 2025-09-07T07:28:18.9508340Z * [new branch] gh/ydwu4/294/orig -> origin/gh/ydwu4/294/orig 2025-09-07T07:28:18.9509627Z * [new branch] gh/ydwu4/295/base -> origin/gh/ydwu4/295/base 2025-09-07T07:28:18.9510601Z * [new branch] gh/ydwu4/295/head -> origin/gh/ydwu4/295/head 2025-09-07T07:28:18.9511416Z * [new branch] gh/ydwu4/295/orig -> origin/gh/ydwu4/295/orig 2025-09-07T07:28:18.9512628Z * [new branch] gh/ydwu4/296/base -> origin/gh/ydwu4/296/base 2025-09-07T07:28:18.9513427Z * [new branch] gh/ydwu4/296/head -> origin/gh/ydwu4/296/head 2025-09-07T07:28:18.9514505Z * [new branch] gh/ydwu4/296/orig -> origin/gh/ydwu4/296/orig 2025-09-07T07:28:18.9516554Z * [new branch] gh/ydwu4/300/base -> origin/gh/ydwu4/300/base 2025-09-07T07:28:18.9517853Z * [new branch] gh/ydwu4/300/head -> origin/gh/ydwu4/300/head 2025-09-07T07:28:18.9518796Z * [new branch] gh/ydwu4/300/orig -> origin/gh/ydwu4/300/orig 2025-09-07T07:28:18.9520318Z * [new branch] gh/ydwu4/301/base -> origin/gh/ydwu4/301/base 2025-09-07T07:28:18.9521172Z * [new branch] gh/ydwu4/301/head -> origin/gh/ydwu4/301/head 2025-09-07T07:28:18.9522317Z * [new branch] gh/ydwu4/301/orig -> origin/gh/ydwu4/301/orig 2025-09-07T07:28:18.9523523Z * [new branch] gh/ydwu4/302/base -> origin/gh/ydwu4/302/base 2025-09-07T07:28:18.9524388Z * [new branch] gh/ydwu4/302/head -> origin/gh/ydwu4/302/head 2025-09-07T07:28:18.9525217Z * [new branch] gh/ydwu4/302/orig -> origin/gh/ydwu4/302/orig 2025-09-07T07:28:18.9526355Z * [new branch] gh/ydwu4/303/base -> origin/gh/ydwu4/303/base 2025-09-07T07:28:18.9527250Z * [new branch] gh/ydwu4/303/head -> origin/gh/ydwu4/303/head 2025-09-07T07:28:18.9528143Z * [new branch] gh/ydwu4/303/orig -> origin/gh/ydwu4/303/orig 2025-09-07T07:28:18.9529208Z * [new branch] gh/ydwu4/304/base -> origin/gh/ydwu4/304/base 2025-09-07T07:28:18.9530091Z * [new branch] gh/ydwu4/304/head -> origin/gh/ydwu4/304/head 2025-09-07T07:28:18.9530966Z * [new branch] gh/ydwu4/304/orig -> origin/gh/ydwu4/304/orig 2025-09-07T07:28:18.9532609Z * [new branch] gh/ydwu4/305/base -> origin/gh/ydwu4/305/base 2025-09-07T07:28:18.9533403Z * [new branch] gh/ydwu4/305/head -> origin/gh/ydwu4/305/head 2025-09-07T07:28:18.9534339Z * [new branch] gh/ydwu4/305/orig -> origin/gh/ydwu4/305/orig 2025-09-07T07:28:18.9535714Z * [new branch] gh/ydwu4/306/base -> origin/gh/ydwu4/306/base 2025-09-07T07:28:18.9536594Z * [new branch] gh/ydwu4/306/head -> origin/gh/ydwu4/306/head 2025-09-07T07:28:18.9537467Z * [new branch] gh/ydwu4/306/orig -> origin/gh/ydwu4/306/orig 2025-09-07T07:28:18.9538665Z * [new branch] gh/ydwu4/307/base -> origin/gh/ydwu4/307/base 2025-09-07T07:28:18.9539469Z * [new branch] gh/ydwu4/307/head -> origin/gh/ydwu4/307/head 2025-09-07T07:28:18.9540330Z * [new branch] gh/ydwu4/307/orig -> origin/gh/ydwu4/307/orig 2025-09-07T07:28:18.9541617Z * [new branch] gh/ydwu4/308/base -> origin/gh/ydwu4/308/base 2025-09-07T07:28:18.9542511Z * [new branch] gh/ydwu4/308/head -> origin/gh/ydwu4/308/head 2025-09-07T07:28:18.9543367Z * [new branch] gh/ydwu4/308/orig -> origin/gh/ydwu4/308/orig 2025-09-07T07:28:18.9544561Z * [new branch] gh/ydwu4/309/base -> origin/gh/ydwu4/309/base 2025-09-07T07:28:18.9545388Z * [new branch] gh/ydwu4/309/head -> origin/gh/ydwu4/309/head 2025-09-07T07:28:18.9546219Z * [new branch] gh/ydwu4/309/orig -> origin/gh/ydwu4/309/orig 2025-09-07T07:28:18.9547524Z * [new branch] gh/ydwu4/310/base -> origin/gh/ydwu4/310/base 2025-09-07T07:28:18.9548551Z * [new branch] gh/ydwu4/310/head -> origin/gh/ydwu4/310/head 2025-09-07T07:28:18.9549404Z * [new branch] gh/ydwu4/310/orig -> origin/gh/ydwu4/310/orig 2025-09-07T07:28:18.9550740Z * [new branch] gh/ydwu4/311/base -> origin/gh/ydwu4/311/base 2025-09-07T07:28:18.9551598Z * [new branch] gh/ydwu4/311/head -> origin/gh/ydwu4/311/head 2025-09-07T07:28:18.9552638Z * [new branch] gh/ydwu4/311/orig -> origin/gh/ydwu4/311/orig 2025-09-07T07:28:18.9553849Z * [new branch] gh/ydwu4/312/base -> origin/gh/ydwu4/312/base 2025-09-07T07:28:18.9554668Z * [new branch] gh/ydwu4/312/head -> origin/gh/ydwu4/312/head 2025-09-07T07:28:18.9555787Z * [new branch] gh/ydwu4/312/orig -> origin/gh/ydwu4/312/orig 2025-09-07T07:28:18.9557172Z * [new branch] gh/ydwu4/313/base -> origin/gh/ydwu4/313/base 2025-09-07T07:28:18.9558200Z * [new branch] gh/ydwu4/313/head -> origin/gh/ydwu4/313/head 2025-09-07T07:28:18.9559020Z * [new branch] gh/ydwu4/313/orig -> origin/gh/ydwu4/313/orig 2025-09-07T07:28:18.9560322Z * [new branch] gh/ydwu4/314/base -> origin/gh/ydwu4/314/base 2025-09-07T07:28:18.9561346Z * [new branch] gh/ydwu4/314/head -> origin/gh/ydwu4/314/head 2025-09-07T07:28:18.9562245Z * [new branch] gh/ydwu4/314/orig -> origin/gh/ydwu4/314/orig 2025-09-07T07:28:18.9563534Z * [new branch] gh/ydwu4/315/base -> origin/gh/ydwu4/315/base 2025-09-07T07:28:18.9564391Z * [new branch] gh/ydwu4/315/head -> origin/gh/ydwu4/315/head 2025-09-07T07:28:18.9565241Z * [new branch] gh/ydwu4/315/orig -> origin/gh/ydwu4/315/orig 2025-09-07T07:28:18.9566682Z * [new branch] gh/ydwu4/316/base -> origin/gh/ydwu4/316/base 2025-09-07T07:28:18.9567602Z * [new branch] gh/ydwu4/316/head -> origin/gh/ydwu4/316/head 2025-09-07T07:28:18.9568470Z * [new branch] gh/ydwu4/316/orig -> origin/gh/ydwu4/316/orig 2025-09-07T07:28:18.9570106Z * [new branch] gh/ydwu4/317/base -> origin/gh/ydwu4/317/base 2025-09-07T07:28:18.9570745Z * [new branch] gh/ydwu4/317/head -> origin/gh/ydwu4/317/head 2025-09-07T07:28:18.9571603Z * [new branch] gh/ydwu4/317/orig -> origin/gh/ydwu4/317/orig 2025-09-07T07:28:18.9572874Z * [new branch] gh/ydwu4/318/base -> origin/gh/ydwu4/318/base 2025-09-07T07:28:18.9573764Z * [new branch] gh/ydwu4/318/head -> origin/gh/ydwu4/318/head 2025-09-07T07:28:18.9574574Z * [new branch] gh/ydwu4/318/orig -> origin/gh/ydwu4/318/orig 2025-09-07T07:28:18.9575755Z * [new branch] gh/ydwu4/319/base -> origin/gh/ydwu4/319/base 2025-09-07T07:28:18.9576609Z * [new branch] gh/ydwu4/319/head -> origin/gh/ydwu4/319/head 2025-09-07T07:28:18.9577488Z * [new branch] gh/ydwu4/319/orig -> origin/gh/ydwu4/319/orig 2025-09-07T07:28:18.9578703Z * [new branch] gh/ydwu4/320/base -> origin/gh/ydwu4/320/base 2025-09-07T07:28:18.9579528Z * [new branch] gh/ydwu4/320/head -> origin/gh/ydwu4/320/head 2025-09-07T07:28:18.9580355Z * [new branch] gh/ydwu4/320/orig -> origin/gh/ydwu4/320/orig 2025-09-07T07:28:18.9581432Z * [new branch] gh/ydwu4/321/base -> origin/gh/ydwu4/321/base 2025-09-07T07:28:18.9582251Z * [new branch] gh/ydwu4/321/head -> origin/gh/ydwu4/321/head 2025-09-07T07:28:18.9583115Z * [new branch] gh/ydwu4/321/orig -> origin/gh/ydwu4/321/orig 2025-09-07T07:28:18.9584299Z * [new branch] gh/ydwu4/322/base -> origin/gh/ydwu4/322/base 2025-09-07T07:28:18.9585158Z * [new branch] gh/ydwu4/322/head -> origin/gh/ydwu4/322/head 2025-09-07T07:28:18.9585953Z * [new branch] gh/ydwu4/322/orig -> origin/gh/ydwu4/322/orig 2025-09-07T07:28:18.9587244Z * [new branch] gh/ydwu4/323/base -> origin/gh/ydwu4/323/base 2025-09-07T07:28:18.9588087Z * [new branch] gh/ydwu4/323/head -> origin/gh/ydwu4/323/head 2025-09-07T07:28:18.9588875Z * [new branch] gh/ydwu4/323/orig -> origin/gh/ydwu4/323/orig 2025-09-07T07:28:18.9590134Z * [new branch] gh/ydwu4/324/base -> origin/gh/ydwu4/324/base 2025-09-07T07:28:18.9591089Z * [new branch] gh/ydwu4/324/head -> origin/gh/ydwu4/324/head 2025-09-07T07:28:18.9591937Z * [new branch] gh/ydwu4/324/orig -> origin/gh/ydwu4/324/orig 2025-09-07T07:28:18.9593689Z * [new branch] gh/yf225/133/base -> origin/gh/yf225/133/base 2025-09-07T07:28:18.9594558Z * [new branch] gh/yf225/133/head -> origin/gh/yf225/133/head 2025-09-07T07:28:18.9596267Z * [new branch] gh/yf225/171/base -> origin/gh/yf225/171/base 2025-09-07T07:28:18.9597252Z * [new branch] gh/yf225/171/head -> origin/gh/yf225/171/head 2025-09-07T07:28:18.9598298Z * [new branch] gh/yf225/171/orig -> origin/gh/yf225/171/orig 2025-09-07T07:28:18.9599699Z * [new branch] gh/yf225/172/base -> origin/gh/yf225/172/base 2025-09-07T07:28:18.9600454Z * [new branch] gh/yf225/172/head -> origin/gh/yf225/172/head 2025-09-07T07:28:18.9601272Z * [new branch] gh/yf225/172/orig -> origin/gh/yf225/172/orig 2025-09-07T07:28:18.9602454Z * [new branch] gh/yf225/93/base -> origin/gh/yf225/93/base 2025-09-07T07:28:18.9603298Z * [new branch] gh/yf225/93/head -> origin/gh/yf225/93/head 2025-09-07T07:28:18.9605265Z * [new branch] gh/yifuwang/152/base -> origin/gh/yifuwang/152/base 2025-09-07T07:28:18.9606576Z * [new branch] gh/yifuwang/152/head -> origin/gh/yifuwang/152/head 2025-09-07T07:28:18.9607584Z * [new branch] gh/yifuwang/152/orig -> origin/gh/yifuwang/152/orig 2025-09-07T07:28:18.9608633Z * [new branch] gh/yifuwang/195/base -> origin/gh/yifuwang/195/base 2025-09-07T07:28:18.9609517Z * [new branch] gh/yifuwang/195/head -> origin/gh/yifuwang/195/head 2025-09-07T07:28:18.9610388Z * [new branch] gh/yifuwang/195/orig -> origin/gh/yifuwang/195/orig 2025-09-07T07:28:18.9611912Z * [new branch] gh/yiming0416/1/base -> origin/gh/yiming0416/1/base 2025-09-07T07:28:18.9612819Z * [new branch] gh/yiming0416/1/head -> origin/gh/yiming0416/1/head 2025-09-07T07:28:18.9613839Z * [new branch] gh/yiming0416/2/base -> origin/gh/yiming0416/2/base 2025-09-07T07:28:18.9614607Z * [new branch] gh/yiming0416/2/head -> origin/gh/yiming0416/2/head 2025-09-07T07:28:18.9616065Z * [new branch] gh/ysiraichi/79/base -> origin/gh/ysiraichi/79/base 2025-09-07T07:28:18.9616924Z * [new branch] gh/ysiraichi/79/head -> origin/gh/ysiraichi/79/head 2025-09-07T07:28:18.9617862Z * [new branch] gh/ysiraichi/79/orig -> origin/gh/ysiraichi/79/orig 2025-09-07T07:28:18.9619022Z * [new branch] gh/ysiraichi/88/base -> origin/gh/ysiraichi/88/base 2025-09-07T07:28:18.9619954Z * [new branch] gh/ysiraichi/88/head -> origin/gh/ysiraichi/88/head 2025-09-07T07:28:18.9620793Z * [new branch] gh/ysiraichi/88/orig -> origin/gh/ysiraichi/88/orig 2025-09-07T07:28:18.9622405Z * [new branch] gh/zhxchen17/25/base -> origin/gh/zhxchen17/25/base 2025-09-07T07:28:18.9623243Z * [new branch] gh/zhxchen17/25/head -> origin/gh/zhxchen17/25/head 2025-09-07T07:28:18.9624086Z * [new branch] gh/zhxchen17/25/orig -> origin/gh/zhxchen17/25/orig 2025-09-07T07:28:18.9625447Z * [new branch] gh/zhxchen17/31/base -> origin/gh/zhxchen17/31/base 2025-09-07T07:28:18.9626352Z * [new branch] gh/zhxchen17/31/head -> origin/gh/zhxchen17/31/head 2025-09-07T07:28:18.9627176Z * [new branch] gh/zhxchen17/31/orig -> origin/gh/zhxchen17/31/orig 2025-09-07T07:28:18.9628362Z * [new branch] gh/zhxchen17/34/base -> origin/gh/zhxchen17/34/base 2025-09-07T07:28:18.9629306Z * [new branch] gh/zhxchen17/34/head -> origin/gh/zhxchen17/34/head 2025-09-07T07:28:18.9630313Z * [new branch] gh/zhxchen17/35/base -> origin/gh/zhxchen17/35/base 2025-09-07T07:28:18.9631224Z * [new branch] gh/zhxchen17/35/head -> origin/gh/zhxchen17/35/head 2025-09-07T07:28:18.9632641Z * [new branch] gh/zhxchen17/37/base -> origin/gh/zhxchen17/37/base 2025-09-07T07:28:18.9633550Z * [new branch] gh/zhxchen17/37/head -> origin/gh/zhxchen17/37/head 2025-09-07T07:28:18.9634473Z * [new branch] gh/zhxchen17/37/orig -> origin/gh/zhxchen17/37/orig 2025-09-07T07:28:18.9635801Z * [new branch] gh/zhxchen17/38/base -> origin/gh/zhxchen17/38/base 2025-09-07T07:28:18.9636568Z * [new branch] gh/zhxchen17/38/head -> origin/gh/zhxchen17/38/head 2025-09-07T07:28:18.9637417Z * [new branch] gh/zhxchen17/38/orig -> origin/gh/zhxchen17/38/orig 2025-09-07T07:28:18.9638515Z * [new branch] gh/zhxchen17/39/base -> origin/gh/zhxchen17/39/base 2025-09-07T07:28:18.9639466Z * [new branch] gh/zhxchen17/39/head -> origin/gh/zhxchen17/39/head 2025-09-07T07:28:18.9640340Z * [new branch] gh/zhxchen17/39/orig -> origin/gh/zhxchen17/39/orig 2025-09-07T07:28:18.9641604Z * [new branch] gh/zhxchen17/40/base -> origin/gh/zhxchen17/40/base 2025-09-07T07:28:18.9642442Z * [new branch] gh/zhxchen17/40/head -> origin/gh/zhxchen17/40/head 2025-09-07T07:28:18.9643527Z * [new branch] gh/zhxchen17/40/orig -> origin/gh/zhxchen17/40/orig 2025-09-07T07:28:18.9644629Z * [new branch] gh/zhxchen17/41/base -> origin/gh/zhxchen17/41/base 2025-09-07T07:28:18.9645941Z * [new branch] gh/zhxchen17/41/head -> origin/gh/zhxchen17/41/head 2025-09-07T07:28:18.9647120Z * [new branch] gh/zhxchen17/41/orig -> origin/gh/zhxchen17/41/orig 2025-09-07T07:28:18.9648491Z * [new branch] gh/zhxchen17/42/base -> origin/gh/zhxchen17/42/base 2025-09-07T07:28:18.9649488Z * [new branch] gh/zhxchen17/42/head -> origin/gh/zhxchen17/42/head 2025-09-07T07:28:18.9650572Z * [new branch] gh/zhxchen17/42/orig -> origin/gh/zhxchen17/42/orig 2025-09-07T07:28:18.9651977Z * [new branch] gh/zhxchen17/43/base -> origin/gh/zhxchen17/43/base 2025-09-07T07:28:18.9652947Z * [new branch] gh/zhxchen17/43/head -> origin/gh/zhxchen17/43/head 2025-09-07T07:28:18.9653890Z * [new branch] gh/zhxchen17/43/orig -> origin/gh/zhxchen17/43/orig 2025-09-07T07:28:18.9655183Z * [new branch] gh/zhxchen17/44/base -> origin/gh/zhxchen17/44/base 2025-09-07T07:28:18.9655957Z * [new branch] gh/zhxchen17/44/head -> origin/gh/zhxchen17/44/head 2025-09-07T07:28:18.9656913Z * [new branch] gh/zhxchen17/44/orig -> origin/gh/zhxchen17/44/orig 2025-09-07T07:28:18.9658049Z * [new branch] gh/zhxchen17/45/base -> origin/gh/zhxchen17/45/base 2025-09-07T07:28:18.9659103Z * [new branch] gh/zhxchen17/45/head -> origin/gh/zhxchen17/45/head 2025-09-07T07:28:18.9659983Z * [new branch] gh/zhxchen17/45/orig -> origin/gh/zhxchen17/45/orig 2025-09-07T07:28:18.9661450Z * [new branch] gh/zklaus/10/base -> origin/gh/zklaus/10/base 2025-09-07T07:28:18.9662329Z * [new branch] gh/zklaus/10/head -> origin/gh/zklaus/10/head 2025-09-07T07:28:18.9663164Z * [new branch] gh/zklaus/10/orig -> origin/gh/zklaus/10/orig 2025-09-07T07:28:18.9664317Z * [new branch] gh/zklaus/11/base -> origin/gh/zklaus/11/base 2025-09-07T07:28:18.9665167Z * [new branch] gh/zklaus/11/head -> origin/gh/zklaus/11/head 2025-09-07T07:28:18.9668929Z * [new branch] gh/zklaus/11/orig -> origin/gh/zklaus/11/orig 2025-09-07T07:28:18.9670102Z * [new branch] gh/zklaus/12/base -> origin/gh/zklaus/12/base 2025-09-07T07:28:18.9671025Z * [new branch] gh/zklaus/12/head -> origin/gh/zklaus/12/head 2025-09-07T07:28:18.9671866Z * [new branch] gh/zklaus/12/orig -> origin/gh/zklaus/12/orig 2025-09-07T07:28:18.9673200Z * [new branch] gh/zklaus/14/base -> origin/gh/zklaus/14/base 2025-09-07T07:28:18.9674030Z * [new branch] gh/zklaus/14/head -> origin/gh/zklaus/14/head 2025-09-07T07:28:18.9674961Z * [new branch] gh/zklaus/14/orig -> origin/gh/zklaus/14/orig 2025-09-07T07:28:18.9676133Z * [new branch] gh/zklaus/15/base -> origin/gh/zklaus/15/base 2025-09-07T07:28:18.9677020Z * [new branch] gh/zklaus/15/head -> origin/gh/zklaus/15/head 2025-09-07T07:28:18.9677957Z * [new branch] gh/zklaus/15/orig -> origin/gh/zklaus/15/orig 2025-09-07T07:28:18.9679188Z * [new branch] gh/zklaus/16/base -> origin/gh/zklaus/16/base 2025-09-07T07:28:18.9680026Z * [new branch] gh/zklaus/16/head -> origin/gh/zklaus/16/head 2025-09-07T07:28:18.9680873Z * [new branch] gh/zklaus/16/orig -> origin/gh/zklaus/16/orig 2025-09-07T07:28:18.9682092Z * [new branch] gh/zklaus/17/base -> origin/gh/zklaus/17/base 2025-09-07T07:28:18.9683060Z * [new branch] gh/zklaus/17/head -> origin/gh/zklaus/17/head 2025-09-07T07:28:18.9683920Z * [new branch] gh/zklaus/17/orig -> origin/gh/zklaus/17/orig 2025-09-07T07:28:18.9685013Z * [new branch] gh/zklaus/18/base -> origin/gh/zklaus/18/base 2025-09-07T07:28:18.9685832Z * [new branch] gh/zklaus/18/head -> origin/gh/zklaus/18/head 2025-09-07T07:28:18.9686692Z * [new branch] gh/zklaus/18/orig -> origin/gh/zklaus/18/orig 2025-09-07T07:28:18.9687789Z * [new branch] gh/zklaus/19/base -> origin/gh/zklaus/19/base 2025-09-07T07:28:18.9688673Z * [new branch] gh/zklaus/19/head -> origin/gh/zklaus/19/head 2025-09-07T07:28:18.9689518Z * [new branch] gh/zklaus/19/orig -> origin/gh/zklaus/19/orig 2025-09-07T07:28:18.9690714Z * [new branch] gh/zklaus/20/base -> origin/gh/zklaus/20/base 2025-09-07T07:28:18.9691571Z * [new branch] gh/zklaus/20/head -> origin/gh/zklaus/20/head 2025-09-07T07:28:18.9692428Z * [new branch] gh/zklaus/20/orig -> origin/gh/zklaus/20/orig 2025-09-07T07:28:18.9693549Z * [new branch] gh/zklaus/7/base -> origin/gh/zklaus/7/base 2025-09-07T07:28:18.9694388Z * [new branch] gh/zklaus/7/head -> origin/gh/zklaus/7/head 2025-09-07T07:28:18.9695287Z * [new branch] gh/zklaus/7/orig -> origin/gh/zklaus/7/orig 2025-09-07T07:28:18.9696398Z * [new branch] gh/zklaus/9/base -> origin/gh/zklaus/9/base 2025-09-07T07:28:18.9697236Z * [new branch] gh/zklaus/9/head -> origin/gh/zklaus/9/head 2025-09-07T07:28:18.9698106Z * [new branch] gh/zklaus/9/orig -> origin/gh/zklaus/9/orig 2025-09-07T07:28:18.9699528Z * [new branch] gh/zou3519/1175/base -> origin/gh/zou3519/1175/base 2025-09-07T07:28:18.9700422Z * [new branch] gh/zou3519/1175/head -> origin/gh/zou3519/1175/head 2025-09-07T07:28:18.9701344Z * [new branch] gh/zou3519/1175/orig -> origin/gh/zou3519/1175/orig 2025-09-07T07:28:18.9702523Z * [new branch] gh/zou3519/1177/base -> origin/gh/zou3519/1177/base 2025-09-07T07:28:18.9703384Z * [new branch] gh/zou3519/1177/head -> origin/gh/zou3519/1177/head 2025-09-07T07:28:18.9704238Z * [new branch] gh/zou3519/1177/orig -> origin/gh/zou3519/1177/orig 2025-09-07T07:28:18.9705467Z * [new branch] gh/zou3519/1191/base -> origin/gh/zou3519/1191/base 2025-09-07T07:28:18.9706486Z * [new branch] gh/zou3519/1191/head -> origin/gh/zou3519/1191/head 2025-09-07T07:28:18.9707362Z * [new branch] gh/zou3519/1191/orig -> origin/gh/zou3519/1191/orig 2025-09-07T07:28:18.9708664Z * [new branch] gh/zou3519/1192/base -> origin/gh/zou3519/1192/base 2025-09-07T07:28:18.9709557Z * [new branch] gh/zou3519/1192/head -> origin/gh/zou3519/1192/head 2025-09-07T07:28:18.9710406Z * [new branch] gh/zou3519/1192/orig -> origin/gh/zou3519/1192/orig 2025-09-07T07:28:18.9711430Z * [new branch] gh/zou3519/1193/base -> origin/gh/zou3519/1193/base 2025-09-07T07:28:18.9712300Z * [new branch] gh/zou3519/1193/head -> origin/gh/zou3519/1193/head 2025-09-07T07:28:18.9713154Z * [new branch] gh/zou3519/1193/orig -> origin/gh/zou3519/1193/orig 2025-09-07T07:28:18.9714166Z * [new branch] gh/zou3519/1194/base -> origin/gh/zou3519/1194/base 2025-09-07T07:28:18.9715062Z * [new branch] gh/zou3519/1194/head -> origin/gh/zou3519/1194/head 2025-09-07T07:28:18.9715889Z * [new branch] gh/zou3519/1194/orig -> origin/gh/zou3519/1194/orig 2025-09-07T07:28:18.9717124Z * [new branch] gh/zou3519/1195/base -> origin/gh/zou3519/1195/base 2025-09-07T07:28:18.9718202Z * [new branch] gh/zou3519/1195/head -> origin/gh/zou3519/1195/head 2025-09-07T07:28:18.9718999Z * [new branch] gh/zou3519/1195/orig -> origin/gh/zou3519/1195/orig 2025-09-07T07:28:18.9720103Z * [new branch] gh/zou3519/1196/base -> origin/gh/zou3519/1196/base 2025-09-07T07:28:18.9721034Z * [new branch] gh/zou3519/1196/head -> origin/gh/zou3519/1196/head 2025-09-07T07:28:18.9721906Z * [new branch] gh/zou3519/1196/orig -> origin/gh/zou3519/1196/orig 2025-09-07T07:28:18.9722947Z * [new branch] gh/zou3519/1197/base -> origin/gh/zou3519/1197/base 2025-09-07T07:28:18.9723826Z * [new branch] gh/zou3519/1197/head -> origin/gh/zou3519/1197/head 2025-09-07T07:28:18.9724684Z * [new branch] gh/zou3519/1197/orig -> origin/gh/zou3519/1197/orig 2025-09-07T07:28:18.9726173Z * [new branch] gh/zpcore/1/base -> origin/gh/zpcore/1/base 2025-09-07T07:28:18.9727054Z * [new branch] gh/zpcore/1/head -> origin/gh/zpcore/1/head 2025-09-07T07:28:18.9728332Z * [new branch] gh/zpcore/10/base -> origin/gh/zpcore/10/base 2025-09-07T07:28:18.9729083Z * [new branch] gh/zpcore/10/head -> origin/gh/zpcore/10/head 2025-09-07T07:28:18.9729936Z * [new branch] gh/zpcore/10/orig -> origin/gh/zpcore/10/orig 2025-09-07T07:28:18.9731247Z * [new branch] gh/zpcore/11/base -> origin/gh/zpcore/11/base 2025-09-07T07:28:18.9732109Z * [new branch] gh/zpcore/11/head -> origin/gh/zpcore/11/head 2025-09-07T07:28:18.9733000Z * [new branch] gh/zpcore/11/orig -> origin/gh/zpcore/11/orig 2025-09-07T07:28:18.9734356Z * [new branch] gh/zpcore/12/base -> origin/gh/zpcore/12/base 2025-09-07T07:28:18.9735451Z * [new branch] gh/zpcore/12/head -> origin/gh/zpcore/12/head 2025-09-07T07:28:18.9736318Z * [new branch] gh/zpcore/12/orig -> origin/gh/zpcore/12/orig 2025-09-07T07:28:18.9737540Z * [new branch] gh/zpcore/13/base -> origin/gh/zpcore/13/base 2025-09-07T07:28:18.9738424Z * [new branch] gh/zpcore/13/head -> origin/gh/zpcore/13/head 2025-09-07T07:28:18.9739273Z * [new branch] gh/zpcore/13/orig -> origin/gh/zpcore/13/orig 2025-09-07T07:28:18.9740495Z * [new branch] gh/zpcore/14/base -> origin/gh/zpcore/14/base 2025-09-07T07:28:18.9741378Z * [new branch] gh/zpcore/14/head -> origin/gh/zpcore/14/head 2025-09-07T07:28:18.9742567Z * [new branch] gh/zpcore/2/base -> origin/gh/zpcore/2/base 2025-09-07T07:28:18.9743449Z * [new branch] gh/zpcore/2/head -> origin/gh/zpcore/2/head 2025-09-07T07:28:18.9744536Z * [new branch] gh/zpcore/3/base -> origin/gh/zpcore/3/base 2025-09-07T07:28:18.9745330Z * [new branch] gh/zpcore/3/head -> origin/gh/zpcore/3/head 2025-09-07T07:28:18.9746347Z * [new branch] gh/zpcore/4/base -> origin/gh/zpcore/4/base 2025-09-07T07:28:18.9747225Z * [new branch] gh/zpcore/4/head -> origin/gh/zpcore/4/head 2025-09-07T07:28:18.9748246Z * [new branch] gh/zpcore/5/base -> origin/gh/zpcore/5/base 2025-09-07T07:28:18.9749063Z * [new branch] gh/zpcore/5/head -> origin/gh/zpcore/5/head 2025-09-07T07:28:18.9750123Z * [new branch] gh/zpcore/6/base -> origin/gh/zpcore/6/base 2025-09-07T07:28:18.9750950Z * [new branch] gh/zpcore/6/head -> origin/gh/zpcore/6/head 2025-09-07T07:28:18.9751991Z * [new branch] gh/zpcore/7/base -> origin/gh/zpcore/7/base 2025-09-07T07:28:18.9752775Z * [new branch] gh/zpcore/7/head -> origin/gh/zpcore/7/head 2025-09-07T07:28:18.9753986Z * [new branch] gh/zpcore/8/base -> origin/gh/zpcore/8/base 2025-09-07T07:28:18.9754781Z * [new branch] gh/zpcore/8/head -> origin/gh/zpcore/8/head 2025-09-07T07:28:18.9755897Z * [new branch] google-main -> origin/google-main 2025-09-07T07:28:18.9757215Z * [new branch] guangyey/external_stream -> origin/guangyey/external_stream 2025-09-07T07:28:18.9758071Z * [new branch] guangyey/host_alloc -> origin/guangyey/host_alloc 2025-09-07T07:28:18.9758737Z * [new branch] guangyey/reimport -> origin/guangyey/reimport 2025-09-07T07:28:18.9759539Z * [new branch] guangyey/test_2025 -> origin/guangyey/test_2025 2025-09-07T07:28:18.9760911Z * [new branch] guilhermeleobas/cherry-pick-55d87d9dfd9 -> origin/guilhermeleobas/cherry-pick-55d87d9dfd9 2025-09-07T07:28:18.9762033Z * [new branch] haozhe/bf16-dynamic-shape -> origin/haozhe/bf16-dynamic-shape 2025-09-07T07:28:18.9763025Z * [new branch] hc_baseline -> origin/hc_baseline 2025-09-07T07:28:18.9764049Z * [new branch] hf_update -> origin/hf_update 2025-09-07T07:28:18.9765003Z * [new branch] hhh_decomp_mul -> origin/hhh_decomp_mul 2025-09-07T07:28:18.9765942Z * [new branch] hhh_rand -> origin/hhh_rand 2025-09-07T07:28:18.9767231Z * [new branch] hoy/mmsplitk -> origin/hoy/mmsplitk 2025-09-07T07:28:18.9768049Z * [new branch] hoy/triton-PR3973 -> origin/hoy/triton-PR3973 2025-09-07T07:28:18.9768969Z * [new branch] hoy/triton-coalescing-baseline -> origin/hoy/triton-coalescing-baseline 2025-09-07T07:28:18.9769766Z * [new branch] hoy/triton-coalescing-new -> origin/hoy/triton-coalescing-new 2025-09-07T07:28:18.9770561Z * [new branch] hoy/triton-coalescing-vec -> origin/hoy/triton-coalescing-vec 2025-09-07T07:28:18.9771487Z * [new branch] inductordecompfix -> origin/inductordecompfix 2025-09-07T07:28:18.9772493Z * [new branch] inline -> origin/inline 2025-09-07T07:28:18.9773345Z * [new branch] inlining -> origin/inlining 2025-09-07T07:28:18.9774329Z * [new branch] inlining-ezyang -> origin/inlining-ezyang 2025-09-07T07:28:18.9775317Z * [new branch] install-torchao-0.13.0 -> origin/install-torchao-0.13.0 2025-09-07T07:28:18.9776190Z * [new branch] int8_sdpa -> origin/int8_sdpa 2025-09-07T07:28:18.9777180Z * [new branch] invoke-subgraph -> origin/invoke-subgraph 2025-09-07T07:28:18.9778102Z * [new branch] issue#58739 -> origin/issue#58739 2025-09-07T07:28:18.9779607Z * [new branch] jcaip/test-cusparselt-version-0.6.2 -> origin/jcaip/test-cusparselt-version-0.6.2 2025-09-07T07:28:18.9780612Z * [new branch] jcaip/update-cusparselt-0.6.2 -> origin/jcaip/update-cusparselt-0.6.2 2025-09-07T07:28:18.9782014Z * [new branch] jeanschmidt/disable_rocm_build_tests -> origin/jeanschmidt/disable_rocm_build_tests 2025-09-07T07:28:18.9783007Z * [new branch] jithunnair-amd-patch-1 -> origin/jithunnair-amd-patch-1 2025-09-07T07:28:18.9784030Z * [new branch] jithunnair-amd-patch-2 -> origin/jithunnair-amd-patch-2 2025-09-07T07:28:18.9785189Z * [new branch] justinchu/attention-tests -> origin/justinchu/attention-tests 2025-09-07T07:28:18.9786040Z * [new branch] justinchu/native-qdq -> origin/justinchu/native-qdq 2025-09-07T07:28:18.9786895Z * [new branch] justinchu/ort-122 -> origin/justinchu/ort-122 2025-09-07T07:28:18.9788124Z * [new branch] justinchuby/dynamo-true -> origin/justinchuby/dynamo-true 2025-09-07T07:28:18.9789496Z * [new branch] kainan666/xlf_debug -> origin/kainan666/xlf_debug 2025-09-07T07:28:18.9790408Z * [new branch] kainan_test -> origin/kainan_test 2025-09-07T07:28:18.9791380Z * [new branch] learnablebias -> origin/learnablebias 2025-09-07T07:28:18.9792702Z * [new branch] leslie/test_group_gemm_epilogues -> origin/leslie/test_group_gemm_epilogues 2025-09-07T07:28:18.9793913Z * [new branch] lessw2020/fix_cutlass_cache_error -> origin/lessw2020/fix_cutlass_cache_error 2025-09-07T07:28:18.9794988Z * [new branch] liaoxuan/shm_all_reduce -> origin/liaoxuan/shm_all_reduce 2025-09-07T07:28:18.9796177Z * [new branch] liaoxuan/test_fa_disable_softmax -> origin/liaoxuan/test_fa_disable_softmax 2025-09-07T07:28:18.9797101Z * [new branch] liaoxuan/test_int8_sdpa -> origin/liaoxuan/test_int8_sdpa 2025-09-07T07:28:18.9798047Z * [new branch] lintbuilddocker -> origin/lintbuilddocker 2025-09-07T07:28:18.9798876Z * [new branch] llama4-stable -> origin/llama4-stable 2025-09-07T07:28:18.9799886Z * [new branch] logdetfix -> origin/logdetfix 2025-09-07T07:28:18.9801491Z * [new branch] lts/release/1.8 -> origin/lts/release/1.8 2025-09-07T07:28:18.9802781Z * [new branch] lucaskabela/#94773 -> origin/lucaskabela/#94773 2025-09-07T07:28:18.9803596Z * [new branch] lucaskabela/flop_counter -> origin/lucaskabela/flop_counter 2025-09-07T07:28:18.9804447Z * [new branch] lucaskabela/func_under_decomp -> origin/lucaskabela/func_under_decomp 2025-09-07T07:28:18.9805328Z * [new branch] lucaskabela/functional_in_dynamo -> origin/lucaskabela/functional_in_dynamo 2025-09-07T07:28:18.9806284Z * [new branch] lucaskabela/install_params_as_graph_attr -> origin/lucaskabela/install_params_as_graph_attr 2025-09-07T07:28:18.9806990Z * [new branch] lucaskabela/issue_120648 -> origin/lucaskabela/issue_120648 2025-09-07T07:28:18.9808097Z * [new branch] lucaskabela/misc_typing_dynamo -> origin/lucaskabela/misc_typing_dynamo 2025-09-07T07:28:18.9809270Z * [new branch] lucaskabela/parameters_as_graph_attr -> origin/lucaskabela/parameters_as_graph_attr 2025-09-07T07:28:18.9810210Z * [new branch] lucaskabela/remove_aot_dispatcher_metadata -> origin/lucaskabela/remove_aot_dispatcher_metadata 2025-09-07T07:28:18.9810976Z * [new branch] lucaskabela/rnn_decomp -> origin/lucaskabela/rnn_decomp 2025-09-07T07:28:18.9811887Z * [new branch] lucaskabela/typing_backends -> origin/lucaskabela/typing_backends 2025-09-07T07:28:18.9812767Z * [new branch] lucaskabela/typing_symbolic_convert -> origin/lucaskabela/typing_symbolic_convert 2025-09-07T07:28:18.9813668Z * [new branch] lucaskabela/typing_utils_improvements -> origin/lucaskabela/typing_utils_improvements 2025-09-07T07:28:18.9814869Z * [new branch] main -> origin/main 2025-09-07T07:28:18.9816011Z * [new branch] main-enable-b200-distributed-tests -> origin/main-enable-b200-distributed-tests 2025-09-07T07:28:18.9816905Z * [new branch] malfet-patch-1 -> origin/malfet-patch-1 2025-09-07T07:28:18.9818003Z * [new branch] malfet-patch-12 -> origin/malfet-patch-12 2025-09-07T07:28:18.9819061Z * [new branch] malfet-patch-14 -> origin/malfet-patch-14 2025-09-07T07:28:18.9820125Z * [new branch] malfet-patch-6 -> origin/malfet-patch-6 2025-09-07T07:28:18.9821119Z * [new branch] malfet-patch-8 -> origin/malfet-patch-8 2025-09-07T07:28:18.9822680Z * [new branch] malfet/be-move-more-settings-to-checkout-pytorch -> origin/malfet/be-move-more-settings-to-checkout-pytorch 2025-09-07T07:28:18.9823470Z * [new branch] malfet/delete-upsteam-cuda -> origin/malfet/delete-upsteam-cuda 2025-09-07T07:28:18.9824239Z * [new branch] malfet/mps-implement-col2im -> origin/malfet/mps-implement-col2im 2025-09-07T07:28:18.9825506Z * [new branch] manuel/test-ops-common-allow-mps -> origin/manuel/test-ops-common-allow-mps 2025-09-07T07:28:18.9826389Z * [new branch] metascroy-patch-1 -> origin/metascroy-patch-1 2025-09-07T07:28:18.9827678Z * [new branch] mlazos/S429861-debug -> origin/mlazos/S429861-debug 2025-09-07T07:28:18.9828517Z * [new branch] mlazos/aa -> origin/mlazos/aa 2025-09-07T07:28:18.9829342Z * [new branch] mlazos/arg-renames -> origin/mlazos/arg-renames 2025-09-07T07:28:18.9830210Z * [new branch] mlazos/backup-test-branch -> origin/mlazos/backup-test-branch 2025-09-07T07:28:18.9831066Z * [new branch] mlazos/bad-cudagraphs -> origin/mlazos/bad-cudagraphs 2025-09-07T07:28:18.9831902Z * [new branch] mlazos/baseline -> origin/mlazos/baseline 2025-09-07T07:28:18.9832769Z * [new branch] mlazos/baseline-graph-breaks -> origin/mlazos/baseline-graph-breaks 2025-09-07T07:28:18.9833522Z * [new branch] mlazos/beta-tensor -> origin/mlazos/beta-tensor 2025-09-07T07:28:18.9834581Z * [new branch] mlazos/better-msg -> origin/mlazos/better-msg 2025-09-07T07:28:18.9835712Z * [new branch] mlazos/buffers -> origin/mlazos/buffers 2025-09-07T07:28:18.9836514Z * [new branch] mlazos/buffers2 -> origin/mlazos/buffers2 2025-09-07T07:28:18.9837389Z * [new branch] mlazos/buffers3 -> origin/mlazos/buffers3 2025-09-07T07:28:18.9838690Z * [new branch] mlazos/ck2 -> origin/mlazos/ck2 2025-09-07T07:28:18.9839698Z * [new branch] mlazos/combokernels -> origin/mlazos/combokernels 2025-09-07T07:28:18.9840571Z * [new branch] mlazos/ctx-cleanup -> origin/mlazos/ctx-cleanup 2025-09-07T07:28:18.9841269Z * [new branch] mlazos/cuda-cmd-log -> origin/mlazos/cuda-cmd-log 2025-09-07T07:28:18.9842342Z * [new branch] mlazos/cudagraph-tests -> origin/mlazos/cudagraph-tests 2025-09-07T07:28:18.9843251Z * [new branch] mlazos/cudagraphs-measurement -> origin/mlazos/cudagraphs-measurement 2025-09-07T07:28:18.9844321Z * [new branch] mlazos/cutlass-test -> origin/mlazos/cutlass-test 2025-09-07T07:28:18.9845253Z * [new branch] mlazos/cutlass-topo-bug -> origin/mlazos/cutlass-topo-bug 2025-09-07T07:28:18.9846124Z * [new branch] mlazos/data-gather -> origin/mlazos/data-gather 2025-09-07T07:28:18.9847067Z * [new branch] mlazos/data-ptrs2 -> origin/mlazos/data-ptrs2 2025-09-07T07:28:18.9847875Z * [new branch] mlazos/data-ptrs3 -> origin/mlazos/data-ptrs3 2025-09-07T07:28:18.9848911Z * [new branch] mlazos/dataclass-proxy -> origin/mlazos/dataclass-proxy 2025-09-07T07:28:18.9849765Z * [new branch] mlazos/dc-attrs -> origin/mlazos/dc-attrs 2025-09-07T07:28:18.9850681Z * [new branch] mlazos/dc-helion -> origin/mlazos/dc-helion 2025-09-07T07:28:18.9851450Z * [new branch] mlazos/dict-fix -> origin/mlazos/dict-fix 2025-09-07T07:28:18.9852566Z * [new branch] mlazos/disable-closures -> origin/mlazos/disable-closures 2025-09-07T07:28:18.9853446Z * [new branch] mlazos/disable-tf -> origin/mlazos/disable-tf 2025-09-07T07:28:18.9854236Z * [new branch] mlazos/dupe-fix -> origin/mlazos/dupe-fix 2025-09-07T07:28:18.9855177Z * [new branch] mlazos/dyn-batch -> origin/mlazos/dyn-batch 2025-09-07T07:28:18.9856147Z * [new branch] mlazos/evt -> origin/mlazos/evt 2025-09-07T07:28:18.9857020Z * [new branch] mlazos/exp_disable -> origin/mlazos/exp_disable 2025-09-07T07:28:18.9857892Z * [new branch] mlazos/extract-examples -> origin/mlazos/extract-examples 2025-09-07T07:28:18.9858795Z * [new branch] mlazos/foreach-op -> origin/mlazos/foreach-op 2025-09-07T07:28:18.9859651Z * [new branch] mlazos/fp8 -> origin/mlazos/fp8 2025-09-07T07:28:18.9860631Z * [new branch] mlazos/fp8-bias -> origin/mlazos/fp8-bias 2025-09-07T07:28:18.9861499Z * [new branch] mlazos/fp8-bias-fusion -> origin/mlazos/fp8-bias-fusion 2025-09-07T07:28:18.9862274Z * [new branch] mlazos/fp8-fixes -> origin/mlazos/fp8-fixes 2025-09-07T07:28:18.9863283Z * [new branch] mlazos/freezing -> origin/mlazos/freezing 2025-09-07T07:28:18.9864210Z * [new branch] mlazos/h-comp -> origin/mlazos/h-comp 2025-09-07T07:28:18.9865055Z * [new branch] mlazos/h-comp2 -> origin/mlazos/h-comp2 2025-09-07T07:28:18.9866064Z * [new branch] mlazos/hash-hop -> origin/mlazos/hash-hop 2025-09-07T07:28:18.9866972Z * [new branch] mlazos/hc -> origin/mlazos/hc 2025-09-07T07:28:18.9867947Z * [new branch] mlazos/hc-cycles -> origin/mlazos/hc-cycles 2025-09-07T07:28:18.9868769Z * [new branch] mlazos/hc-fixes -> origin/mlazos/hc-fixes 2025-09-07T07:28:18.9869687Z * [new branch] mlazos/hc-fixes3 -> origin/mlazos/hc-fixes3 2025-09-07T07:28:18.9870570Z * [new branch] mlazos/hc-fixes4 -> origin/mlazos/hc-fixes4 2025-09-07T07:28:18.9871430Z * [new branch] mlazos/hc-hf -> origin/mlazos/hc-hf 2025-09-07T07:28:18.9872300Z * [new branch] mlazos/hc-mut -> origin/mlazos/hc-mut 2025-09-07T07:28:18.9873190Z * [new branch] mlazos/hc10 -> origin/mlazos/hc10 2025-09-07T07:28:18.9874035Z * [new branch] mlazos/hc11 -> origin/mlazos/hc11 2025-09-07T07:28:18.9874964Z * [new branch] mlazos/hc12 -> origin/mlazos/hc12 2025-09-07T07:28:18.9875951Z * [new branch] mlazos/hc13 -> origin/mlazos/hc13 2025-09-07T07:28:18.9876804Z * [new branch] mlazos/hc14 -> origin/mlazos/hc14 2025-09-07T07:28:18.9877703Z * [new branch] mlazos/hc15 -> origin/mlazos/hc15 2025-09-07T07:28:18.9878692Z * [new branch] mlazos/hc2 -> origin/mlazos/hc2 2025-09-07T07:28:18.9879522Z * [new branch] mlazos/hc4 -> origin/mlazos/hc4 2025-09-07T07:28:18.9880419Z * [new branch] mlazos/hc5 -> origin/mlazos/hc5 2025-09-07T07:28:18.9881260Z * [new branch] mlazos/hc6 -> origin/mlazos/hc6 2025-09-07T07:28:18.9882165Z * [new branch] mlazos/hc7 -> origin/mlazos/hc7 2025-09-07T07:28:18.9882944Z * [new branch] mlazos/hc8 -> origin/mlazos/hc8 2025-09-07T07:28:18.9883823Z * [new branch] mlazos/hc9 -> origin/mlazos/hc9 2025-09-07T07:28:18.9884731Z * [new branch] mlazos/hc_baseline2 -> origin/mlazos/hc_baseline2 2025-09-07T07:28:18.9885626Z * [new branch] mlazos/init-per-param -> origin/mlazos/init-per-param 2025-09-07T07:28:18.9886480Z * [new branch] mlazos/init_per_param -> origin/mlazos/init_per_param 2025-09-07T07:28:18.9887418Z * [new branch] mlazos/less-guards -> origin/mlazos/less-guards 2025-09-07T07:28:18.9888316Z * [new branch] mlazos/lr-composibility -> origin/mlazos/lr-composibility 2025-09-07T07:28:18.9889297Z * [new branch] mlazos/main -> origin/mlazos/main 2025-09-07T07:28:18.9890110Z * [new branch] mlazos/main-test-enablement -> origin/mlazos/main-test-enablement 2025-09-07T07:28:18.9890914Z * [new branch] mlazos/main2 -> origin/mlazos/main2 2025-09-07T07:28:18.9891884Z * [new branch] mlazos/mark-static-update -> origin/mlazos/mark-static-update 2025-09-07T07:28:18.9892645Z * [new branch] mlazos/mcg -> origin/mlazos/mcg 2025-09-07T07:28:18.9893535Z * [new branch] mlazos/mcg2 -> origin/mlazos/mcg2 2025-09-07T07:28:18.9894444Z * [new branch] mlazos/meta-guards -> origin/mlazos/meta-guards 2025-09-07T07:28:18.9895534Z * [new branch] mlazos/mlazos/ck2 -> origin/mlazos/mlazos/ck2 2025-09-07T07:28:18.9896510Z * [new branch] mlazos/mlazos/foreach-map-adam -> origin/mlazos/mlazos/foreach-map-adam 2025-09-07T07:28:18.9897591Z * [new branch] mlazos/mlazos/tf-mode-backup -> origin/mlazos/mlazos/tf-mode-backup 2025-09-07T07:28:18.9898486Z * [new branch] mlazos/mod-fix -> origin/mlazos/mod-fix 2025-09-07T07:28:18.9899456Z * [new branch] mlazos/mode-fix -> origin/mlazos/mode-fix 2025-09-07T07:28:18.9900362Z * [new branch] mlazos/more-tests -> origin/mlazos/more-tests 2025-09-07T07:28:18.9901206Z * [new branch] mlazos/no-cpp -> origin/mlazos/no-cpp 2025-09-07T07:28:18.9902239Z * [new branch] mlazos/no-init-group-handling -> origin/mlazos/no-init-group-handling 2025-09-07T07:28:18.9903238Z * [new branch] mlazos/offsets -> origin/mlazos/offsets 2025-09-07T07:28:18.9904168Z * [new branch] mlazos/opt-bench-exp2 -> origin/mlazos/opt-bench-exp2 2025-09-07T07:28:18.9905234Z * [new branch] mlazos/opt-incr -> origin/mlazos/opt-incr 2025-09-07T07:28:18.9906192Z * [new branch] mlazos/proxy-ctors -> origin/mlazos/proxy-ctors 2025-09-07T07:28:18.9907068Z * [new branch] mlazos/quant-fix -> origin/mlazos/quant-fix 2025-09-07T07:28:18.9907983Z * [new branch] mlazos/resnet-fix -> origin/mlazos/resnet-fix 2025-09-07T07:28:18.9908886Z * [new branch] mlazos/revert-inline -> origin/mlazos/revert-inline 2025-09-07T07:28:18.9909827Z * [new branch] mlazos/rm-buf-names -> origin/mlazos/rm-buf-names 2025-09-07T07:28:18.9910589Z * [new branch] mlazos/rm-code -> origin/mlazos/rm-code 2025-09-07T07:28:18.9911485Z * [new branch] mlazos/rm-spam -> origin/mlazos/rm-spam 2025-09-07T07:28:18.9912389Z * [new branch] mlazos/rtp -> origin/mlazos/rtp 2025-09-07T07:28:18.9913476Z * [new branch] mlazos/static-idx-dbg -> origin/mlazos/static-idx-dbg 2025-09-07T07:28:18.9914403Z * [new branch] mlazos/static-inputs-log -> origin/mlazos/static-inputs-log 2025-09-07T07:28:18.9915267Z * [new branch] mlazos/sub-param-fix -> origin/mlazos/sub-param-fix 2025-09-07T07:28:18.9916149Z * [new branch] mlazos/td-fix2 -> origin/mlazos/td-fix2 2025-09-07T07:28:18.9917057Z * [new branch] mlazos/tensor-hasattr2 -> origin/mlazos/tensor-hasattr2 2025-09-07T07:28:18.9917953Z * [new branch] mlazos/test -> origin/mlazos/test 2025-09-07T07:28:18.9918865Z * [new branch] mlazos/tf-mode -> origin/mlazos/tf-mode 2025-09-07T07:28:18.9919831Z * [new branch] mlazos/tf-mode-backup2 -> origin/mlazos/tf-mode-backup2 2025-09-07T07:28:18.9920714Z * [new branch] mlazos/tf-mode-reland -> origin/mlazos/tf-mode-reland 2025-09-07T07:28:18.9921710Z * [new branch] mlazos/tf-mode-reland2 -> origin/mlazos/tf-mode-reland2 2025-09-07T07:28:18.9922665Z * [new branch] mlazos/tf-mode-reland3 -> origin/mlazos/tf-mode-reland3 2025-09-07T07:28:18.9923467Z * [new branch] mlazos/topo-fix -> origin/mlazos/topo-fix 2025-09-07T07:28:18.9924384Z * [new branch] mlazos/triton-no-epi -> origin/mlazos/triton-no-epi 2025-09-07T07:28:18.9925317Z * [new branch] mlazos/tune-proto -> origin/mlazos/tune-proto 2025-09-07T07:28:18.9926129Z * [new branch] mlazos/tuple-fixes -> origin/mlazos/tuple-fixes 2025-09-07T07:28:18.9926857Z * [new branch] mlazos/tuple-fixes2 -> origin/mlazos/tuple-fixes2 2025-09-07T07:28:18.9927811Z * [new branch] mlazos/tuple-handling -> origin/mlazos/tuple-handling 2025-09-07T07:28:18.9928692Z * [new branch] mlazos/user-streams -> origin/mlazos/user-streams 2025-09-07T07:28:18.9929669Z * [new branch] mlazos/vary-beta -> origin/mlazos/vary-beta 2025-09-07T07:28:18.9930590Z * [new branch] mlazos/vary-beta2 -> origin/mlazos/vary-beta2 2025-09-07T07:28:18.9931437Z * [new branch] mlazos/weird-perf1 -> origin/mlazos/weird-perf1 2025-09-07T07:28:18.9932563Z * [new branch] mm_out_dtype_compile -> origin/mm_out_dtype_compile 2025-09-07T07:28:18.9933522Z * [new branch] modify-setupvllm -> origin/modify-setupvllm 2025-09-07T07:28:18.9934504Z * [new branch] module-shim -> origin/module-shim 2025-09-07T07:28:18.9935506Z * [new branch] move-theme-out-docker -> origin/move-theme-out-docker 2025-09-07T07:28:18.9936793Z * [new branch] msaroufim/be1 -> origin/msaroufim/be1 2025-09-07T07:28:18.9937683Z * [new branch] msaroufim/cn_path -> origin/msaroufim/cn_path 2025-09-07T07:28:18.9938618Z * [new branch] msaroufim/dtensorfusedadam -> origin/msaroufim/dtensorfusedadam 2025-09-07T07:28:18.9939450Z * [new branch] msaroufim/reduce -> origin/msaroufim/reduce 2025-09-07T07:28:18.9940672Z * [new branch] mtia/basic-cmake -> origin/mtia/basic-cmake 2025-09-07T07:28:18.9941661Z * [new branch] muon_dev -> origin/muon_dev 2025-09-07T07:28:18.9942621Z * [new branch] muon_dev_1 -> origin/muon_dev_1 2025-09-07T07:28:18.9943588Z * [new branch] nativert_num_outputs -> origin/nativert_num_outputs 2025-09-07T07:28:18.9944698Z * [new branch] nativert_numoutputs -> origin/nativert_numoutputs 2025-09-07T07:28:18.9945655Z * [new branch] new-modifiy-setupvllm -> origin/new-modifiy-setupvllm 2025-09-07T07:28:18.9946618Z * [new branch] new-setupvllm -> origin/new-setupvllm 2025-09-07T07:28:18.9947596Z * [new branch] new_zeros_dtype -> origin/new_zeros_dtype 2025-09-07T07:28:18.9948585Z * [new branch] newtest-base -> origin/newtest-base 2025-09-07T07:28:18.9949790Z * [new branch] ngimel/cat_perf1 -> origin/ngimel/cat_perf1 2025-09-07T07:28:18.9950583Z * [new branch] ngimel/einsum_fix -> origin/ngimel/einsum_fix 2025-09-07T07:28:18.9951445Z * [new branch] ngimel/error_index_list -> origin/ngimel/error_index_list 2025-09-07T07:28:18.9952212Z * [new branch] ngimel/fabric_check -> origin/ngimel/fabric_check 2025-09-07T07:28:18.9952962Z * [new branch] ngimel/fabric_fix -> origin/ngimel/fabric_fix 2025-09-07T07:28:18.9953904Z * [new branch] ngimel/fix_driver_init_error -> origin/ngimel/fix_driver_init_error 2025-09-07T07:28:18.9954967Z * [new branch] ngimel/fix_nccl_segment_seg -> origin/ngimel/fix_nccl_segment_seg 2025-09-07T07:28:18.9956052Z * [new branch] ngimel/gg_new -> origin/ngimel/gg_new 2025-09-07T07:28:18.9956966Z * [new branch] ngimel/modeguard -> origin/ngimel/modeguard 2025-09-07T07:28:18.9958151Z * [new branch] ngimel/multicast_fix -> origin/ngimel/multicast_fix 2025-09-07T07:28:18.9959107Z * [new branch] ngimel/rocm_handle_type -> origin/ngimel/rocm_handle_type 2025-09-07T07:28:18.9959972Z * [new branch] ngimel/symm_handle_fabric -> origin/ngimel/symm_handle_fabric 2025-09-07T07:28:18.9960830Z * [new branch] ngimel/unbind_multimem -> origin/ngimel/unbind_multimem 2025-09-07T07:28:18.9961826Z * [new branch] nightly -> origin/nightly 2025-09-07T07:28:18.9962886Z * [new branch] nmacchioni-patch-10 -> origin/nmacchioni-patch-10 2025-09-07T07:28:18.9963947Z * [new branch] nmacchioni-patch-7 -> origin/nmacchioni-patch-7 2025-09-07T07:28:18.9965014Z * [new branch] nmacchioni-patch-8 -> origin/nmacchioni-patch-8 2025-09-07T07:28:18.9966122Z * [new branch] nmacchioni-patch-9 -> origin/nmacchioni-patch-9 2025-09-07T07:28:18.9967448Z * [new branch] nullplay/fuse_matmul -> origin/nullplay/fuse_matmul 2025-09-07T07:28:18.9968403Z * [new branch] nullplay_fuse_matmul -> origin/nullplay_fuse_matmul 2025-09-07T07:28:18.9969311Z * [new branch] one-off -> origin/one-off 2025-09-07T07:28:18.9970943Z * [new branch] orig/release/1.10 -> origin/orig/release/1.10 2025-09-07T07:28:18.9971905Z * [new branch] orig/release/1.11 -> origin/orig/release/1.11 2025-09-07T07:28:18.9972866Z * [new branch] orig/release/1.12 -> origin/orig/release/1.12 2025-09-07T07:28:18.9973975Z * [new branch] orig/release/1.13 -> origin/orig/release/1.13 2025-09-07T07:28:18.9975055Z * [new branch] orig/release/1.6 -> origin/orig/release/1.6 2025-09-07T07:28:18.9976065Z * [new branch] orig/release/1.7 -> origin/orig/release/1.7 2025-09-07T07:28:18.9977055Z * [new branch] orig/release/1.8 -> origin/orig/release/1.8 2025-09-07T07:28:18.9978195Z * [new branch] orig/release/1.9 -> origin/orig/release/1.9 2025-09-07T07:28:18.9979141Z * [new branch] orig/release/2.0 -> origin/orig/release/2.0 2025-09-07T07:28:18.9980042Z * [new branch] orig/release/2.1 -> origin/orig/release/2.1 2025-09-07T07:28:18.9980919Z * [new branch] orig/release/2.2 -> origin/orig/release/2.2 2025-09-07T07:28:18.9981843Z * [new branch] orig/release/2.3 -> origin/orig/release/2.3 2025-09-07T07:28:18.9982717Z * [new branch] orig/release/2.4 -> origin/orig/release/2.4 2025-09-07T07:28:18.9983627Z * [new branch] orig/release/2.5 -> origin/orig/release/2.5 2025-09-07T07:28:18.9984467Z * [new branch] orig/release/2.6 -> origin/orig/release/2.6 2025-09-07T07:28:18.9985655Z * [new branch] orig/release/2.7 -> origin/orig/release/2.7 2025-09-07T07:28:18.9986947Z * [new branch] orig/release/2.8 -> origin/orig/release/2.8 2025-09-07T07:28:18.9988183Z * [new branch] oulgen/fx_graph -> origin/oulgen/fx_graph 2025-09-07T07:28:18.9989180Z * [new branch] padded-tensor -> origin/padded-tensor 2025-09-07T07:28:18.9990298Z * [new branch] pca2 -> origin/pca2 2025-09-07T07:28:18.9991428Z * [new branch] pianpwk-patch-1 -> origin/pianpwk-patch-1 2025-09-07T07:28:18.9992771Z * [new branch] pianpwk/backed_size_oblivious_export -> origin/pianpwk/backed_size_oblivious_export 2025-09-07T07:28:18.9993740Z * [new branch] pianpwk/invalidate_fake_memo -> origin/pianpwk/invalidate_fake_memo 2025-09-07T07:28:18.9994424Z * [new branch] pianpwk/max_1_strides -> origin/pianpwk/max_1_strides 2025-09-07T07:28:18.9995221Z * [new branch] pianpwk/maybe_guard_rel -> origin/pianpwk/maybe_guard_rel 2025-09-07T07:28:18.9996049Z * [new branch] pianpwk/nonzero_memo -> origin/pianpwk/nonzero_memo 2025-09-07T07:28:18.9997181Z * [new branch] pianpwk/oblivious_reshape_view_better -> origin/pianpwk/oblivious_reshape_view_better 2025-09-07T07:28:18.9998286Z * [new branch] pianpwk/oblivious_slice_forward -> origin/pianpwk/oblivious_slice_forward 2025-09-07T07:28:18.9999133Z * [new branch] pianpwk/oblivious_where -> origin/pianpwk/oblivious_where 2025-09-07T07:28:19.0000040Z * [new branch] pianpwk/param_static_pgo -> origin/pianpwk/param_static_pgo 2025-09-07T07:28:19.0001661Z * [new branch] pianpwk/pre_forward_hook -> origin/pianpwk/pre_forward_hook 2025-09-07T07:28:19.0001982Z * [new branch] pianpwk/remove_guard_fail_break -> origin/pianpwk/remove_guard_fail_break 2025-09-07T07:28:19.0002865Z * [new branch] pianpwk/slice_fresh_symbols -> origin/pianpwk/slice_fresh_symbols 2025-09-07T07:28:19.0003499Z * [new branch] pianpwk/sym_tokens_draft -> origin/pianpwk/sym_tokens_draft 2025-09-07T07:28:19.0004541Z * [new branch] pianpwk/test_pointwise_guard_or_false -> origin/pianpwk/test_pointwise_guard_or_false 2025-09-07T07:28:19.0005309Z * [new branch] pianpwk/test_slice_fake_impl -> origin/pianpwk/test_slice_fake_impl 2025-09-07T07:28:19.0006203Z * [new branch] pianpwk/totally_draft_sym_wrap -> origin/pianpwk/totally_draft_sym_wrap 2025-09-07T07:28:19.0007064Z * [new branch] pianpwk/unbacked_channels_last -> origin/pianpwk/unbacked_channels_last 2025-09-07T07:28:19.0007934Z * [new branch] pianpwk/unbacked_safe_conv1d -> origin/pianpwk/unbacked_safe_conv1d 2025-09-07T07:28:19.0008775Z * [new branch] pianpwk/unbacked_sdpa_flash -> origin/pianpwk/unbacked_sdpa_flash 2025-09-07T07:28:19.0009726Z * [new branch] pianpwk/unbacked_should_swap -> origin/pianpwk/unbacked_should_swap 2025-09-07T07:28:19.0010558Z * [new branch] pianpwk/unbacked_should_swap_2 -> origin/pianpwk/unbacked_should_swap_2 2025-09-07T07:28:19.0011409Z * [new branch] pianpwk/unbacked_slice_binding -> origin/pianpwk/unbacked_slice_binding 2025-09-07T07:28:19.0012295Z * [new branch] pianpwk/unbacked_slice_forward -> origin/pianpwk/unbacked_slice_forward 2025-09-07T07:28:19.0013079Z * [new branch] pianpwk/user_symints -> origin/pianpwk/user_symints 2025-09-07T07:28:19.0013979Z * [new branch] pianpwk/wan21_reshape -> origin/pianpwk/wan21_reshape 2025-09-07T07:28:19.0014870Z * [new branch] pianpwk/whitelist_optimizer -> origin/pianpwk/whitelist_optimizer 2025-09-07T07:28:19.0015900Z * [new branch] pin-torchao -> origin/pin-torchao 2025-09-07T07:28:19.0017195Z * [new branch] piz/fall_back_missing_0716 -> origin/piz/fall_back_missing_0716 2025-09-07T07:28:19.0018085Z * [new branch] piz/improve_scatter_0808 -> origin/piz/improve_scatter_0808 2025-09-07T07:28:19.0018978Z * [new branch] pool-separate -> origin/pool-separate 2025-09-07T07:28:19.0019971Z * [new branch] pr-156087 -> origin/pr-156087 2025-09-07T07:28:19.0021264Z * [new branch] pr/131860 -> origin/pr/131860 2025-09-07T07:28:19.0022247Z * [new branch] predispatch_to -> origin/predispatch_to 2025-09-07T07:28:19.0023213Z * [new branch] pt-opt-cuda3 -> origin/pt-opt-cuda3 2025-09-07T07:28:19.0024169Z * [new branch] pyobjectslot -> origin/pyobjectslot 2025-09-07T07:28:19.0025477Z * [new branch] python_compiled_autograd -> origin/python_compiled_autograd 2025-09-07T07:28:19.0027032Z * [new branch] qchip/export-D54134695 -> origin/qchip/export-D54134695 2025-09-07T07:28:19.0027975Z * [new branch] quint-bits -> origin/quint-bits 2025-09-07T07:28:19.0029378Z * [new branch] release/1.10 -> origin/release/1.10 2025-09-07T07:28:19.0030397Z * [new branch] release/1.11 -> origin/release/1.11 2025-09-07T07:28:19.0031293Z * [new branch] release/1.12 -> origin/release/1.12 2025-09-07T07:28:19.0032285Z * [new branch] release/1.13 -> origin/release/1.13 2025-09-07T07:28:19.0033115Z * [new branch] release/1.4 -> origin/release/1.4 2025-09-07T07:28:19.0033859Z * [new branch] release/1.4.1 -> origin/release/1.4.1 2025-09-07T07:28:19.0034792Z * [new branch] release/1.5 -> origin/release/1.5 2025-09-07T07:28:19.0035764Z * [new branch] release/1.6 -> origin/release/1.6 2025-09-07T07:28:19.0036693Z * [new branch] release/1.7 -> origin/release/1.7 2025-09-07T07:28:19.0037695Z * [new branch] release/1.8 -> origin/release/1.8 2025-09-07T07:28:19.0038565Z * [new branch] release/1.9 -> origin/release/1.9 2025-09-07T07:28:19.0039458Z * [new branch] release/2.0 -> origin/release/2.0 2025-09-07T07:28:19.0040444Z * [new branch] release/2.1 -> origin/release/2.1 2025-09-07T07:28:19.0041466Z * [new branch] release/2.2 -> origin/release/2.2 2025-09-07T07:28:19.0042679Z * [new branch] release/2.3 -> origin/release/2.3 2025-09-07T07:28:19.0043902Z * [new branch] release/2.4 -> origin/release/2.4 2025-09-07T07:28:19.0045100Z * [new branch] release/2.5 -> origin/release/2.5 2025-09-07T07:28:19.0046139Z * [new branch] release/2.6 -> origin/release/2.6 2025-09-07T07:28:19.0047148Z * [new branch] release/2.7 -> origin/release/2.7 2025-09-07T07:28:19.0048176Z * [new branch] release/2.8 -> origin/release/2.8 2025-09-07T07:28:19.0049153Z * [new branch] release_notes -> origin/release_notes 2025-09-07T07:28:19.0050161Z * [new branch] remove-actionable-label -> origin/remove-actionable-label 2025-09-07T07:28:19.0051063Z * [new branch] remove-ao -> origin/remove-ao 2025-09-07T07:28:19.0052302Z * [new branch] removedeprecatedvllmtest -> origin/removedeprecatedvllmtest 2025-09-07T07:28:19.0053262Z * [new branch] replace-pytorch-labs-20250812-195836 -> origin/replace-pytorch-labs-20250812-195836 2025-09-07T07:28:19.0054116Z * [new branch] replace-pytorch-labs-20250812-200248 -> origin/replace-pytorch-labs-20250812-200248 2025-09-07T07:28:19.0054913Z * [new branch] replace-pytorch-labs-20250812-200324 -> origin/replace-pytorch-labs-20250812-200324 2025-09-07T07:28:19.0055952Z * [new branch] replace-pytorch-labs-20250812-204020 -> origin/replace-pytorch-labs-20250812-204020 2025-09-07T07:28:19.0056845Z * [new branch] replace-pytorch-labs-20250812-204125 -> origin/replace-pytorch-labs-20250812-204125 2025-09-07T07:28:19.0057836Z * [new branch] replace-pytorch-labs-20250812-205624 -> origin/replace-pytorch-labs-20250812-205624 2025-09-07T07:28:19.0059702Z * [new branch] revert-131069-gh/krzysztofjordan/1/head -> origin/revert-131069-gh/krzysztofjordan/1/head 2025-09-07T07:28:19.0061686Z * [new branch] revert-131469-gh/andrewor14/51/head -> origin/revert-131469-gh/andrewor14/51/head 2025-09-07T07:28:19.0063604Z * [new branch] revert-156870-gh/skarjala/3/head -> origin/revert-156870-gh/skarjala/3/head 2025-09-07T07:28:19.0064752Z * [new branch] revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ -> origin/revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ 2025-09-07T07:28:19.0065563Z * [new branch] rocm-monitoring -> origin/rocm-monitoring 2025-09-07T07:28:19.0069703Z * [new branch] ruisi/relax_memory -> origin/ruisi/relax_memory 2025-09-07T07:28:19.0070683Z * [new branch] run-torchbench-smoke-test-h100 -> origin/run-torchbench-smoke-test-h100 2025-09-07T07:28:19.0072069Z * [new branch] ryanguo99/cleanup-dynamo-expected-failures -> origin/ryanguo99/cleanup-dynamo-expected-failures 2025-09-07T07:28:19.0072850Z * [new branch] ryanguo99/fix-closure-var -> origin/ryanguo99/fix-closure-var 2025-09-07T07:28:19.0074221Z * [new branch] rzou/faketensor_bench -> origin/rzou/faketensor_bench 2025-09-07T07:28:19.0075049Z * [new branch] rzou/njt -> origin/rzou/njt 2025-09-07T07:28:19.0076297Z * [new branch] rzou/pca -> origin/rzou/pca 2025-09-07T07:28:19.0077186Z * [new branch] rzou/realprop -> origin/rzou/realprop 2025-09-07T07:28:19.0078177Z * [new branch] rzou/setup_context -> origin/rzou/setup_context 2025-09-07T07:28:19.0079662Z * [new branch] sanchitintel/refactor_aten_int8_woq_gemm -> origin/sanchitintel/refactor_aten_int8_woq_gemm 2025-09-07T07:28:19.0080763Z * [new branch] sanchitintel/weird_thing_with_test_cpu_select_algorithm -> origin/sanchitintel/weird_thing_with_test_cpu_select_algorithm 2025-09-07T07:28:19.0081685Z * [new branch] sapling-pr-archive-SS-JIA -> origin/sapling-pr-archive-SS-JIA 2025-09-07T07:28:19.0082598Z * [new branch] save -> origin/save 2025-09-07T07:28:19.0083840Z * [new branch] sdym/2.5.1 -> origin/sdym/2.5.1 2025-09-07T07:28:19.0084936Z * [new branch] seemethere-patch-1 -> origin/seemethere-patch-1 2025-09-07T07:28:19.0085892Z * [new branch] setupvllm -> origin/setupvllm 2025-09-07T07:28:19.0086891Z * [new branch] share_and_pin_fork -> origin/share_and_pin_fork 2025-09-07T07:28:19.0088188Z * [new branch] shengf/fx-xform-perf -> origin/shengf/fx-xform-perf 2025-09-07T07:28:19.0089177Z * [new branch] shikaili_fp8_allgather -> origin/shikaili_fp8_allgather 2025-09-07T07:28:19.0090397Z * [new branch] shoumikhin-patch-1 -> origin/shoumikhin-patch-1 2025-09-07T07:28:19.0091436Z * [new branch] shoumikhin-patch-12 -> origin/shoumikhin-patch-12 2025-09-07T07:28:19.0092451Z * [new branch] simplify-fq-per-channel -> origin/simplify-fq-per-channel 2025-09-07T07:28:19.0093454Z * [new branch] solve-accuracy-fix -> origin/solve-accuracy-fix 2025-09-07T07:28:19.0094538Z * [new branch] soulitzer/stash-tls-ac -> origin/soulitzer/stash-tls-ac 2025-09-07T07:28:19.0095875Z * [new branch] sqzhang/flight4 -> origin/sqzhang/flight4 2025-09-07T07:28:19.0096788Z * [new branch] sqzhang/flight4plus -> origin/sqzhang/flight4plus 2025-09-07T07:28:19.0098019Z * [new branch] sraikund/record_funct_test -> origin/sraikund/record_funct_test 2025-09-07T07:28:19.0099229Z * [new branch] sraikund16/test -> origin/sraikund16/test 2025-09-07T07:28:19.0100273Z * [new branch] stablize-compilation-time -> origin/stablize-compilation-time 2025-09-07T07:28:19.0101307Z * [new branch] standalone-templates -> origin/standalone-templates 2025-09-07T07:28:19.0102670Z * [new branch] standalone_package_weights -> origin/standalone_package_weights 2025-09-07T07:28:19.0103514Z * [new branch] starterTaskUpdate -> origin/starterTaskUpdate 2025-09-07T07:28:19.0104476Z * [new branch] subgraph_fuse -> origin/subgraph_fuse 2025-09-07T07:28:19.0105566Z * [new branch] support-uv-in-collect_env -> origin/support-uv-in-collect_env 2025-09-07T07:28:19.0106511Z * [new branch] sve-poc -> origin/sve-poc 2025-09-07T07:28:19.0107444Z * [new branch] svekars-patch-1 -> origin/svekars-patch-1 2025-09-07T07:28:19.0108418Z * [new branch] switch-bn -> origin/switch-bn 2025-09-07T07:28:19.0109537Z * [new branch] sympy-bottleneck-repro -> origin/sympy-bottleneck-repro 2025-09-07T07:28:19.0110797Z * [new branch] tenpercent/ck_rocm_ci_v3 -> origin/tenpercent/ck_rocm_ci_v3 2025-09-07T07:28:19.0111844Z * [new branch] tensordict_integration -> origin/tensordict_integration 2025-09-07T07:28:19.0112772Z * [new branch] test-7054 -> origin/test-7054 2025-09-07T07:28:19.0113881Z * [new branch] test-move-conda-builds -> origin/test-move-conda-builds 2025-09-07T07:28:19.0114881Z * [new branch] test-myst-markdown-docstring -> origin/test-myst-markdown-docstring 2025-09-07T07:28:19.0115740Z * [new branch] test-old -> origin/test-old 2025-09-07T07:28:19.0116869Z * [new branch] test-vec-migration-internally -> origin/test-vec-migration-internally 2025-09-07T07:28:19.0118003Z * [new branch] test/bmm_heur -> origin/test/bmm_heur 2025-09-07T07:28:19.0118899Z * [new branch] test/inductor -> origin/test/inductor 2025-09-07T07:28:19.0120198Z * [new branch] tianren/flex_paged_attn_fix -> origin/tianren/flex_paged_attn_fix 2025-09-07T07:28:19.0121079Z * [new branch] tianren/flex_paged_attn_fix_temp -> origin/tianren/flex_paged_attn_fix_temp 2025-09-07T07:28:19.0121776Z * [new branch] tianren/test -> origin/tianren/test 2025-09-07T07:28:19.0122749Z * [new branch] tidy_performance_cyy -> origin/tidy_performance_cyy 2025-09-07T07:28:19.0123681Z * [new branch] torchtitan_ep -> origin/torchtitan_ep 2025-09-07T07:28:19.0124738Z * [new branch] trace_fsdp_torchtune_lora -> origin/trace_fsdp_torchtune_lora 2025-09-07T07:28:19.0125699Z * [new branch] traceable_fsdp_unit_tests -> origin/traceable_fsdp_unit_tests 2025-09-07T07:28:19.0126663Z * [new branch] tree_loop_vec_base -> origin/tree_loop_vec_base 2025-09-07T07:28:19.0127609Z * [new branch] tree_vec_base -> origin/tree_vec_base 2025-09-07T07:28:19.0128639Z * [new branch] triton-update -> origin/triton-update 2025-09-07T07:28:19.0129627Z * [new branch] triton_kernel -> origin/triton_kernel 2025-09-07T07:28:19.0130544Z * [new branch] triton_kernel_perf -> origin/triton_kernel_perf 2025-09-07T07:28:19.0131465Z * [new branch] tt_pkg_1908 -> origin/tt_pkg_1908 2025-09-07T07:28:19.0132542Z * [new branch] tweak-transformer-dependabot -> origin/tweak-transformer-dependabot 2025-09-07T07:28:19.0133400Z * [new branch] type_dec -> origin/type_dec 2025-09-07T07:28:19.0134525Z * [new branch] udate-sphinx-dependancies -> origin/udate-sphinx-dependancies 2025-09-07T07:28:19.0135913Z * [new branch] update-audio-commit-hash/16818882925-1712-1 -> origin/update-audio-commit-hash/16818882925-1712-1 2025-09-07T07:28:19.0136807Z * [new branch] update-audio-commit-hash/16895560422-1720-1 -> origin/update-audio-commit-hash/16895560422-1720-1 2025-09-07T07:28:19.0137777Z * [new branch] update-audio-commit-hash/16924174496-1738-1 -> origin/update-audio-commit-hash/16924174496-1738-1 2025-09-07T07:28:19.0138489Z * [new branch] update-audio-commit-hash/17002010821-1749-1 -> origin/update-audio-commit-hash/17002010821-1749-1 2025-09-07T07:28:19.0139342Z * [new branch] update-audio-commit-hash/17056004427-1766-1 -> origin/update-audio-commit-hash/17056004427-1766-1 2025-09-07T07:28:19.0140349Z * [new branch] update-audio-commit-hash/17085054029-1767-1 -> origin/update-audio-commit-hash/17085054029-1767-1 2025-09-07T07:28:19.0141439Z * [new branch] update-audio-commit-hash/17142507405-1771-1 -> origin/update-audio-commit-hash/17142507405-1771-1 2025-09-07T07:28:19.0142518Z * [new branch] update-audio-commit-hash/17168762740-1773-1 -> origin/update-audio-commit-hash/17168762740-1773-1 2025-09-07T07:28:19.0143404Z * [new branch] update-audio-commit-hash/17311174639-1780-1 -> origin/update-audio-commit-hash/17311174639-1780-1 2025-09-07T07:28:19.0144317Z * [new branch] update-audio-commit-hash/17336898740-1781-1 -> origin/update-audio-commit-hash/17336898740-1781-1 2025-09-07T07:28:19.0145180Z * [new branch] update-audio-commit-hash/17389727684-1786-1 -> origin/update-audio-commit-hash/17389727684-1786-1 2025-09-07T07:28:19.0146039Z * [new branch] update-audio-commit-hash/17449538142-1790-1 -> origin/update-audio-commit-hash/17449538142-1790-1 2025-09-07T07:28:19.0146959Z * [new branch] update-audio-commit-hash/17507351808-1794-1 -> origin/update-audio-commit-hash/17507351808-1794-1 2025-09-07T07:28:19.0147819Z * [new branch] update-dynamic-shapes-doc -> origin/update-dynamic-shapes-doc 2025-09-07T07:28:19.0149183Z * [new branch] update-executorch-commit-hash/15694981040-1626-1 -> origin/update-executorch-commit-hash/15694981040-1626-1 2025-09-07T07:28:19.0150567Z * [new branch] update-triton-commit-hash/13663274526-1487-2 -> origin/update-triton-commit-hash/13663274526-1487-2 2025-09-07T07:28:19.0151786Z * [new branch] update-vision-commit-hash/15336342773-1607-1 -> origin/update-vision-commit-hash/15336342773-1607-1 2025-09-07T07:28:19.0153007Z * [new branch] update-vllm-commit-hash/16737365217-1704-1 -> origin/update-vllm-commit-hash/16737365217-1704-1 2025-09-07T07:28:19.0153815Z * [new branch] update-vllm-commit-hash/16843157111-1713-1 -> origin/update-vllm-commit-hash/16843157111-1713-1 2025-09-07T07:28:19.0154690Z * [new branch] update-vllm-commit-hash/16855312394-1714-1 -> origin/update-vllm-commit-hash/16855312394-1714-1 2025-09-07T07:28:19.0155520Z * [new branch] update-vllm-commit-hash/16924174496-1738-1 -> origin/update-vllm-commit-hash/16924174496-1738-1 2025-09-07T07:28:19.0156288Z * [new branch] update-vllm-commit-hash/16952608705-1745-1 -> origin/update-vllm-commit-hash/16952608705-1745-1 2025-09-07T07:28:19.0157401Z * [new branch] update-vllm-commit-hash/16979836546-1748-1 -> origin/update-vllm-commit-hash/16979836546-1748-1 2025-09-07T07:28:19.0158462Z * [new branch] update-vllm-commit-hash/17014576881-1756-1 -> origin/update-vllm-commit-hash/17014576881-1756-1 2025-09-07T07:28:19.0159492Z * [new branch] update-vllm-commit-hash/17027830869-1761-1 -> origin/update-vllm-commit-hash/17027830869-1761-1 2025-09-07T07:28:19.0160447Z * [new branch] update-vllm-commit-hash/17056004427-1766-1 -> origin/update-vllm-commit-hash/17056004427-1766-1 2025-09-07T07:28:19.0161297Z * [new branch] update-vllm-commit-hash/17085054029-1767-1 -> origin/update-vllm-commit-hash/17085054029-1767-1 2025-09-07T07:28:19.0162214Z * [new branch] update-vllm-commit-hash/17113610216-1768-1 -> origin/update-vllm-commit-hash/17113610216-1768-1 2025-09-07T07:28:19.0163188Z * [new branch] update-vllm-commit-hash/17142507405-1771-1 -> origin/update-vllm-commit-hash/17142507405-1771-1 2025-09-07T07:28:19.0164063Z * [new branch] update-vllm-commit-hash/17181878974-1774-1 -> origin/update-vllm-commit-hash/17181878974-1774-1 2025-09-07T07:28:19.0164949Z * [new branch] update-vllm-commit-hash/17311174639-1780-1 -> origin/update-vllm-commit-hash/17311174639-1780-1 2025-09-07T07:28:19.0165844Z * [new branch] update-vllm-commit-hash/17336898740-1781-1 -> origin/update-vllm-commit-hash/17336898740-1781-1 2025-09-07T07:28:19.0166754Z * [new branch] update-vllm-commit-hash/17364352302-1785-1 -> origin/update-vllm-commit-hash/17364352302-1785-1 2025-09-07T07:28:19.0167584Z * [new branch] update-vllm-commit-hash/17389727684-1786-1 -> origin/update-vllm-commit-hash/17389727684-1786-1 2025-09-07T07:28:19.0168521Z * [new branch] update-vllm-commit-hash/17449538142-1790-1 -> origin/update-vllm-commit-hash/17449538142-1790-1 2025-09-07T07:28:19.0169414Z * [new branch] update-vllm-commit-hash/17480069797-1791-1 -> origin/update-vllm-commit-hash/17480069797-1791-1 2025-09-07T07:28:19.0170258Z * [new branch] update-vllm-commit-hash/17507351808-1794-1 -> origin/update-vllm-commit-hash/17507351808-1794-1 2025-09-07T07:28:19.0171619Z * [new branch] update-xla-commit-hash/16873912760-198-1 -> origin/update-xla-commit-hash/16873912760-198-1 2025-09-07T07:28:19.0172460Z * [new branch] update-xla-commit-hash/17034266655-199-1 -> origin/update-xla-commit-hash/17034266655-199-1 2025-09-07T07:28:19.0173222Z * [new branch] update-xla-commit-hash/17202464405-200-1 -> origin/update-xla-commit-hash/17202464405-200-1 2025-09-07T07:28:19.0174299Z * [new branch] update_docs_torch_multinomial_issue#125388 -> origin/update_docs_torch_multinomial_issue#125388 2025-09-07T07:28:19.0175136Z * [new branch] update_executorch_pin -> origin/update_executorch_pin 2025-09-07T07:28:19.0176128Z * [new branch] update_slow_tests_1722488736 -> origin/update_slow_tests_1722488736 2025-09-07T07:28:19.0177078Z * [new branch] update_slow_tests_1722879173 -> origin/update_slow_tests_1722879173 2025-09-07T07:28:19.0178099Z * [new branch] update_slow_tests_1752478971 -> origin/update_slow_tests_1752478971 2025-09-07T07:28:19.0179031Z * [new branch] update_slow_tests_1755502951 -> origin/update_slow_tests_1755502951 2025-09-07T07:28:19.0179991Z * [new branch] update_slow_tests_1756107664 -> origin/update_slow_tests_1756107664 2025-09-07T07:28:19.0181007Z * [new branch] update_submodule_FBGEMM -> origin/update_submodule_FBGEMM 2025-09-07T07:28:19.0181893Z * [new branch] update_submodule_kineto -> origin/update_submodule_kineto 2025-09-07T07:28:19.0182859Z * [new branch] update_submodule_tensorpipe -> origin/update_submodule_tensorpipe 2025-09-07T07:28:19.0183881Z * [new branch] v0.1.2 -> origin/v0.1.2 2025-09-07T07:28:19.0184968Z * [new branch] v1.0.1 -> origin/v1.0.1 2025-09-07T07:28:19.0186059Z * [new branch] v1.0.3 -> origin/v1.0.3 2025-09-07T07:28:19.0187104Z * [new branch] v1.1.0 -> origin/v1.1.0 2025-09-07T07:28:19.0188204Z * [new branch] v1.2.0 -> origin/v1.2.0 2025-09-07T07:28:19.0189177Z * [new branch] v1.3.0 -> origin/v1.3.0 2025-09-07T07:28:19.0190237Z * [new branch] v1.3.1 -> origin/v1.3.1 2025-09-07T07:28:19.0191285Z * [new branch] validate_fn -> origin/validate_fn 2025-09-07T07:28:19.0192497Z * [new branch] validations_2.6 -> origin/validations_2.6 2025-09-07T07:28:19.0193647Z * [new branch] validations_2.8 -> origin/validations_2.8 2025-09-07T07:28:19.0194968Z * [new branch] viable/strict -> origin/viable/strict 2025-09-07T07:28:19.0195847Z * [new branch] vllmbuildci -> origin/vllmbuildci 2025-09-07T07:28:19.0196820Z * [new branch] vllmpin -> origin/vllmpin 2025-09-07T07:28:19.0198111Z * [new branch] wdvr/conda_devcontainer -> origin/wdvr/conda_devcontainer 2025-09-07T07:28:19.0198910Z * [new branch] wdvr/iss_145259 -> origin/wdvr/iss_145259 2025-09-07T07:28:19.0200047Z * [new branch] weight_sharing_cpp -> origin/weight_sharing_cpp 2025-09-07T07:28:19.0201689Z * [new branch] whc/flight4 -> origin/whc/flight4 2025-09-07T07:28:19.0202629Z * [new branch] whc/flight51 -> origin/whc/flight51 2025-09-07T07:28:19.0203438Z * [new branch] whc/flight53 -> origin/whc/flight53 2025-09-07T07:28:19.0204367Z * [new branch] whc/stage2 -> origin/whc/stage2 2025-09-07T07:28:19.0205222Z * [new branch] whc/uneven -> origin/whc/uneven 2025-09-07T07:28:19.0206431Z * [new branch] whc/uneven-merge -> origin/whc/uneven-merge 2025-09-07T07:28:19.0207416Z * [new branch] win_warnings -> origin/win_warnings 2025-09-07T07:28:19.0208376Z * [new branch] windows_libtorch_free -> origin/windows_libtorch_free 2025-09-07T07:28:19.0209167Z * [new branch] workonoldcommit -> origin/workonoldcommit 2025-09-07T07:28:19.0210470Z * [new branch] wychi-autotune-prune-configs-by-shared-mem -> origin/wychi-autotune-prune-configs-by-shared-mem 2025-09-07T07:28:19.0211469Z * [new branch] xmfan/ca_0516 -> origin/xmfan/ca_0516 2025-09-07T07:28:19.0212326Z * [new branch] xmfan/ca_1051b93192 -> origin/xmfan/ca_1051b93192 2025-09-07T07:28:19.0213439Z * [new branch] xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 -> origin/xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 2025-09-07T07:28:19.0214128Z * [new branch] xmfan/ca_5a2be192d1 -> origin/xmfan/ca_5a2be192d1 2025-09-07T07:28:19.0214939Z * [new branch] xmfan/ca_9d59b516e9 -> origin/xmfan/ca_9d59b516e9 2025-09-07T07:28:19.0215890Z * [new branch] xmfan/ca_api -> origin/xmfan/ca_api 2025-09-07T07:28:19.0216641Z * [new branch] xmfan/ca_apr8 -> origin/xmfan/ca_apr8 2025-09-07T07:28:19.0217728Z * [new branch] xmfan/ca_base -> origin/xmfan/ca_base 2025-09-07T07:28:19.0219219Z * [new branch] xmfan/ca_cudagraphs -> origin/xmfan/ca_cudagraphs 2025-09-07T07:28:19.0220148Z * [new branch] xmfan/ca_dynamic -> origin/xmfan/ca_dynamic 2025-09-07T07:28:19.0221085Z * [new branch] xmfan/ca_fix_dyn -> origin/xmfan/ca_fix_dyn 2025-09-07T07:28:19.0221954Z * [new branch] xmfan/ca_fix_lowering -> origin/xmfan/ca_fix_lowering 2025-09-07T07:28:19.0222796Z * [new branch] xmfan/ca_fix_polyfills -> origin/xmfan/ca_fix_polyfills 2025-09-07T07:28:19.0223518Z * [new branch] xmfan/ca_jan3 -> origin/xmfan/ca_jan3 2025-09-07T07:28:19.0224456Z * [new branch] xmfan/ca_jun18 -> origin/xmfan/ca_jun18 2025-09-07T07:28:19.0225375Z * [new branch] xmfan/ca_jun24 -> origin/xmfan/ca_jun24 2025-09-07T07:28:19.0226241Z * [new branch] xmfan/ca_mem_base -> origin/xmfan/ca_mem_base 2025-09-07T07:28:19.0227053Z * [new branch] xmfan/ca_mem_fix -> origin/xmfan/ca_mem_fix 2025-09-07T07:28:19.0227947Z * [new branch] xmfan/ca_memory_fix -> origin/xmfan/ca_memory_fix 2025-09-07T07:28:19.0228962Z * [new branch] xmfan/ca_memory_fix_rebased -> origin/xmfan/ca_memory_fix_rebased 2025-09-07T07:28:19.0229957Z * [new branch] xmfan/ca_memory_fix_rebased2 -> origin/xmfan/ca_memory_fix_rebased2 2025-09-07T07:28:19.0230759Z * [new branch] xmfan/ca_move_to_cuda -> origin/xmfan/ca_move_to_cuda 2025-09-07T07:28:19.0231617Z * [new branch] xmfan/ca_nested -> origin/xmfan/ca_nested 2025-09-07T07:28:19.0232515Z * [new branch] xmfan/ca_overhead -> origin/xmfan/ca_overhead 2025-09-07T07:28:19.0233426Z * [new branch] xmfan/ca_overhead_0eba7e5451 -> origin/xmfan/ca_overhead_0eba7e5451 2025-09-07T07:28:19.0234173Z * [new branch] xmfan/ca_scalar -> origin/xmfan/ca_scalar 2025-09-07T07:28:19.0235132Z * [new branch] xmfan/ca_subclass_mem_fix -> origin/xmfan/ca_subclass_mem_fix 2025-09-07T07:28:19.0235957Z * [new branch] xmfan/ca_warm_mem -> origin/xmfan/ca_warm_mem 2025-09-07T07:28:19.0236778Z * [new branch] xmfan/ca_warm_mem_base -> origin/xmfan/ca_warm_mem_base 2025-09-07T07:28:19.0237605Z * [new branch] xmfan/cacu_jun18 -> origin/xmfan/cacu_jun18 2025-09-07T07:28:19.0238511Z * [new branch] xmfan/cacu_jun19 -> origin/xmfan/cacu_jun19 2025-09-07T07:28:19.0239331Z * [new branch] xmfan/cacu_jun4 -> origin/xmfan/cacu_jun4 2025-09-07T07:28:19.0240268Z * [new branch] xmfan/cacu_may27 -> origin/xmfan/cacu_may27 2025-09-07T07:28:19.0241171Z * [new branch] xmfan/disable_duck_shape -> origin/xmfan/disable_duck_shape 2025-09-07T07:28:19.0242077Z * [new branch] xmfan/fca_cpp_node_passthrough -> origin/xmfan/fca_cpp_node_passthrough 2025-09-07T07:28:19.0242857Z * [new branch] xmfan/issue_123374 -> origin/xmfan/issue_123374 2025-09-07T07:28:19.0243933Z * [new branch] xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 2025-09-07T07:28:19.0244878Z * [new branch] xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 2025-09-07T07:28:19.0245549Z * [new branch] xmfan/segfault_test -> origin/xmfan/segfault_test 2025-09-07T07:28:19.0246429Z * [new branch] xmfan/single_step -> origin/xmfan/single_step 2025-09-07T07:28:19.0247294Z * [new branch] xmfan/sth_0829 -> origin/xmfan/sth_0829 2025-09-07T07:28:19.0248196Z * [new branch] xmfan/test -> origin/xmfan/test 2025-09-07T07:28:19.0249506Z * [new branch] yguo/debug-0226-constexpr -> origin/yguo/debug-0226-constexpr 2025-09-07T07:28:19.0250547Z * [new branch] yguo/new_latest_changes -> origin/yguo/new_latest_changes 2025-09-07T07:28:19.0251388Z * [new branch] yguo/patch_constexpr_changes -> origin/yguo/patch_constexpr_changes 2025-09-07T07:28:19.0252367Z * [new branch] yihan_quantization -> origin/yihan_quantization 2025-09-07T07:28:19.0253710Z * [new branch] yiming/add_jit_trace_benchmark -> origin/yiming/add_jit_trace_benchmark 2025-09-07T07:28:19.0254750Z * [new branch] yiming/add_nativert_benchmark -> origin/yiming/add_nativert_benchmark 2025-09-07T07:28:19.0255533Z * [new branch] yiming/bootcamp -> origin/yiming/bootcamp 2025-09-07T07:28:19.0256762Z * [new branch] zainr/canary-test -> origin/zainr/canary-test 2025-09-07T07:28:19.0257827Z * [new branch] zainr/cleanup-gh-runners -> origin/zainr/cleanup-gh-runners 2025-09-07T07:28:19.0258627Z * [new branch] zainr/git-push-v2 -> origin/zainr/git-push-v2 2025-09-07T07:28:19.0259447Z * [new branch] zainr/pull-migration-c -> origin/zainr/pull-migration-c 2025-09-07T07:28:19.0260275Z * [new branch] zainr/test -> origin/zainr/test 2025-09-07T07:28:19.0261219Z * [new branch] zainr/test2 -> origin/zainr/test2 2025-09-07T07:28:19.0261916Z * [new branch] zainr/unstable -> origin/zainr/unstable 2025-09-07T07:28:19.0262733Z * [new branch] zainr/unstable-xla -> origin/zainr/unstable-xla 2025-09-07T07:28:19.0263935Z * [new branch] zasdfgbnm-patch-3 -> origin/zasdfgbnm-patch-3 2025-09-07T07:28:19.0264826Z * [new branch] zb2p -> origin/zb2p 2025-09-07T07:28:19.0265975Z * [new branch] zero_grad_optimization -> origin/zero_grad_optimization 2025-09-07T07:28:19.0266945Z * [new branch] zeros-and-scatter-part2 -> origin/zeros-and-scatter-part2 2025-09-07T07:28:19.0268406Z * [new branch] zhxchen17/scratch/0 -> origin/zhxchen17/scratch/0 2025-09-07T07:28:19.0269743Z * [new branch] zhxhcen17/moodycamel -> origin/zhxhcen17/moodycamel 2025-09-07T07:28:19.0271045Z * [new branch] zxiiro/main -> origin/zxiiro/main 2025-09-07T07:28:19.0272157Z * [new tag] bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug -> bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug 2025-09-07T07:28:19.0272769Z * [new tag] ci/binaries/77164 -> ci/binaries/77164 2025-09-07T07:28:19.0273563Z * [new tag] ciflow/binaries/156049 -> ciflow/binaries/156049 2025-09-07T07:28:19.0274107Z * [new tag] ciflow/binaries/156712 -> ciflow/binaries/156712 2025-09-07T07:28:19.0274588Z * [new tag] ciflow/binaries/157432 -> ciflow/binaries/157432 2025-09-07T07:28:19.0275165Z * [new tag] ciflow/binaries/157685 -> ciflow/binaries/157685 2025-09-07T07:28:19.0275747Z * [new tag] ciflow/binaries/157689 -> ciflow/binaries/157689 2025-09-07T07:28:19.0276242Z * [new tag] ciflow/binaries/158104 -> ciflow/binaries/158104 2025-09-07T07:28:19.0276921Z * [new tag] ciflow/binaries/160229 -> ciflow/binaries/160229 2025-09-07T07:28:19.0277610Z * [new tag] ciflow/binaries/160720 -> ciflow/binaries/160720 2025-09-07T07:28:19.0278111Z * [new tag] ciflow/binaries/162080 -> ciflow/binaries/162080 2025-09-07T07:28:19.0278689Z * [new tag] ciflow/binaries/162329 -> ciflow/binaries/162329 2025-09-07T07:28:19.0279345Z * [new tag] ciflow/binaries_libtorch/156049 -> ciflow/binaries_libtorch/156049 2025-09-07T07:28:19.0279876Z * [new tag] ciflow/binaries_libtorch/156711 -> ciflow/binaries_libtorch/156711 2025-09-07T07:28:19.0280437Z * [new tag] ciflow/binaries_libtorch/157432 -> ciflow/binaries_libtorch/157432 2025-09-07T07:28:19.0281058Z * [new tag] ciflow/binaries_wheel/156049 -> ciflow/binaries_wheel/156049 2025-09-07T07:28:19.0281586Z * [new tag] ciflow/binaries_wheel/156711 -> ciflow/binaries_wheel/156711 2025-09-07T07:28:19.0282033Z * [new tag] ciflow/binaries_wheel/157432 -> ciflow/binaries_wheel/157432 2025-09-07T07:28:19.0282550Z * [new tag] ciflow/binaries_wheel/162136 -> ciflow/binaries_wheel/162136 2025-09-07T07:28:19.0283192Z * [new tag] ciflow/binaries_wheel/162252 -> ciflow/binaries_wheel/162252 2025-09-07T07:28:19.0283720Z * [new tag] ciflow/binaries_wheel/162325 -> ciflow/binaries_wheel/162325 2025-09-07T07:28:19.0284512Z * [new tag] ciflow/h100-distributed/156703 -> ciflow/h100-distributed/156703 2025-09-07T07:28:19.0285211Z * [new tag] ciflow/h100-symm-mem/157635 -> ciflow/h100-symm-mem/157635 2025-09-07T07:28:19.0285726Z * [new tag] ciflow/h100-symm-mem/161984 -> ciflow/h100-symm-mem/161984 2025-09-07T07:28:19.0286257Z * [new tag] ciflow/h100-symm-mem/162003 -> ciflow/h100-symm-mem/162003 2025-09-07T07:28:19.0286893Z * [new tag] ciflow/h100-symm-mem/162011 -> ciflow/h100-symm-mem/162011 2025-09-07T07:28:19.0287320Z * [new tag] ciflow/h100-symm-mem/162026 -> ciflow/h100-symm-mem/162026 2025-09-07T07:28:19.0287844Z * [new tag] ciflow/h100-symm-mem/162033 -> ciflow/h100-symm-mem/162033 2025-09-07T07:28:19.0288354Z * [new tag] ciflow/h100-symm-mem/162040 -> ciflow/h100-symm-mem/162040 2025-09-07T07:28:19.0288891Z * [new tag] ciflow/h100-symm-mem/162041 -> ciflow/h100-symm-mem/162041 2025-09-07T07:28:19.0289410Z * [new tag] ciflow/h100-symm-mem/162142 -> ciflow/h100-symm-mem/162142 2025-09-07T07:28:19.0289929Z * [new tag] ciflow/h100-symm-mem/162150 -> ciflow/h100-symm-mem/162150 2025-09-07T07:28:19.0290458Z * [new tag] ciflow/h100-symm-mem/162243 -> ciflow/h100-symm-mem/162243 2025-09-07T07:28:19.0291232Z * [new tag] ciflow/h100-symm-mem/162320 -> ciflow/h100-symm-mem/162320 2025-09-07T07:28:19.0291924Z * [new tag] ciflow/h100/159158 -> ciflow/h100/159158 2025-09-07T07:28:19.0292919Z * [new tag] ciflow/h100/160480 -> ciflow/h100/160480 2025-09-07T07:28:19.0293536Z * [new tag] ciflow/h100/161749 -> ciflow/h100/161749 2025-09-07T07:28:19.0294150Z * [new tag] ciflow/h100/162022 -> ciflow/h100/162022 2025-09-07T07:28:19.0294727Z * [new tag] ciflow/h100/162278 -> ciflow/h100/162278 2025-09-07T07:28:19.0295673Z * [new tag] ciflow/inductor-perf-test-nightly-rocm/156592 -> ciflow/inductor-perf-test-nightly-rocm/156592 2025-09-07T07:28:19.0296485Z * [new tag] ciflow/inductor-perf-test-nightly/156592 -> ciflow/inductor-perf-test-nightly/156592 2025-09-07T07:28:19.0297137Z * [new tag] ciflow/inductor-periodic/162063 -> ciflow/inductor-periodic/162063 2025-09-07T07:28:19.0297703Z * [new tag] ciflow/inductor-periodic/162227 -> ciflow/inductor-periodic/162227 2025-09-07T07:28:19.0298323Z * [new tag] ciflow/inductor-periodic/162323 -> ciflow/inductor-periodic/162323 2025-09-07T07:28:19.0299030Z * [new tag] ciflow/inductor-rocm/154170 -> ciflow/inductor-rocm/154170 2025-09-07T07:28:19.0299777Z * [new tag] ciflow/inductor-rocm/159146 -> ciflow/inductor-rocm/159146 2025-09-07T07:28:19.0300301Z * [new tag] ciflow/inductor-rocm/159158 -> ciflow/inductor-rocm/159158 2025-09-07T07:28:19.0300951Z * [new tag] ciflow/inductor-rocm/161715 -> ciflow/inductor-rocm/161715 2025-09-07T07:28:19.0301699Z * [new tag] ciflow/inductor-rocm/162053 -> ciflow/inductor-rocm/162053 2025-09-07T07:28:19.0302419Z * [new tag] ciflow/inductor-rocm/162056 -> ciflow/inductor-rocm/162056 2025-09-07T07:28:19.0303031Z * [new tag] ciflow/inductor/137400 -> ciflow/inductor/137400 2025-09-07T07:28:19.0303569Z * [new tag] ciflow/inductor/148180 -> ciflow/inductor/148180 2025-09-07T07:28:19.0304077Z * [new tag] ciflow/inductor/148328 -> ciflow/inductor/148328 2025-09-07T07:28:19.0304646Z * [new tag] ciflow/inductor/148484 -> ciflow/inductor/148484 2025-09-07T07:28:19.0305135Z * [new tag] ciflow/inductor/148492 -> ciflow/inductor/148492 2025-09-07T07:28:19.0305679Z * [new tag] ciflow/inductor/152624 -> ciflow/inductor/152624 2025-09-07T07:28:19.0306215Z * [new tag] ciflow/inductor/154694 -> ciflow/inductor/154694 2025-09-07T07:28:19.0306737Z * [new tag] ciflow/inductor/156049 -> ciflow/inductor/156049 2025-09-07T07:28:19.0307281Z * [new tag] ciflow/inductor/156592 -> ciflow/inductor/156592 2025-09-07T07:28:19.0307759Z * [new tag] ciflow/inductor/157635 -> ciflow/inductor/157635 2025-09-07T07:28:19.0308364Z * [new tag] ciflow/inductor/157685 -> ciflow/inductor/157685 2025-09-07T07:28:19.0309075Z * [new tag] ciflow/inductor/157686 -> ciflow/inductor/157686 2025-09-07T07:28:19.0309804Z * [new tag] ciflow/inductor/157689 -> ciflow/inductor/157689 2025-09-07T07:28:19.0310453Z * [new tag] ciflow/inductor/157699 -> ciflow/inductor/157699 2025-09-07T07:28:19.0311190Z * [new tag] ciflow/inductor/157743 -> ciflow/inductor/157743 2025-09-07T07:28:19.0311877Z * [new tag] ciflow/inductor/157994 -> ciflow/inductor/157994 2025-09-07T07:28:19.0312451Z * [new tag] ciflow/inductor/158091 -> ciflow/inductor/158091 2025-09-07T07:28:19.0312997Z * [new tag] ciflow/inductor/158104 -> ciflow/inductor/158104 2025-09-07T07:28:19.0313723Z * [new tag] ciflow/inductor/158404 -> ciflow/inductor/158404 2025-09-07T07:28:19.0314283Z * [new tag] ciflow/inductor/158647 -> ciflow/inductor/158647 2025-09-07T07:28:19.0315027Z * [new tag] ciflow/inductor/158932 -> ciflow/inductor/158932 2025-09-07T07:28:19.0315577Z * [new tag] ciflow/inductor/159146 -> ciflow/inductor/159146 2025-09-07T07:28:19.0316126Z * [new tag] ciflow/inductor/159158 -> ciflow/inductor/159158 2025-09-07T07:28:19.0316881Z * [new tag] ciflow/inductor/159274 -> ciflow/inductor/159274 2025-09-07T07:28:19.0317434Z * [new tag] ciflow/inductor/159664 -> ciflow/inductor/159664 2025-09-07T07:28:19.0318165Z * [new tag] ciflow/inductor/159778 -> ciflow/inductor/159778 2025-09-07T07:28:19.0318716Z * [new tag] ciflow/inductor/159835 -> ciflow/inductor/159835 2025-09-07T07:28:19.0319404Z * [new tag] ciflow/inductor/159944 -> ciflow/inductor/159944 2025-09-07T07:28:19.0320145Z * [new tag] ciflow/inductor/160161 -> ciflow/inductor/160161 2025-09-07T07:28:19.0320772Z * [new tag] ciflow/inductor/160174 -> ciflow/inductor/160174 2025-09-07T07:28:19.0321462Z * [new tag] ciflow/inductor/160323 -> ciflow/inductor/160323 2025-09-07T07:28:19.0322267Z * [new tag] ciflow/inductor/160324 -> ciflow/inductor/160324 2025-09-07T07:28:19.0323037Z * [new tag] ciflow/inductor/160325 -> ciflow/inductor/160325 2025-09-07T07:28:19.0323764Z * [new tag] ciflow/inductor/160326 -> ciflow/inductor/160326 2025-09-07T07:28:19.0324403Z * [new tag] ciflow/inductor/160327 -> ciflow/inductor/160327 2025-09-07T07:28:19.0325011Z * [new tag] ciflow/inductor/160328 -> ciflow/inductor/160328 2025-09-07T07:28:19.0325675Z * [new tag] ciflow/inductor/160329 -> ciflow/inductor/160329 2025-09-07T07:28:19.0326224Z * [new tag] ciflow/inductor/160480 -> ciflow/inductor/160480 2025-09-07T07:28:19.0326996Z * [new tag] ciflow/inductor/160532 -> ciflow/inductor/160532 2025-09-07T07:28:19.0328153Z * [new tag] ciflow/inductor/160539 -> ciflow/inductor/160539 2025-09-07T07:28:19.0328739Z * [new tag] ciflow/inductor/160580 -> ciflow/inductor/160580 2025-09-07T07:28:19.0329301Z * [new tag] ciflow/inductor/160685 -> ciflow/inductor/160685 2025-09-07T07:28:19.0329829Z * [new tag] ciflow/inductor/160686 -> ciflow/inductor/160686 2025-09-07T07:28:19.0330425Z * [new tag] ciflow/inductor/160687 -> ciflow/inductor/160687 2025-09-07T07:28:19.0330975Z * [new tag] ciflow/inductor/160688 -> ciflow/inductor/160688 2025-09-07T07:28:19.0331526Z * [new tag] ciflow/inductor/160690 -> ciflow/inductor/160690 2025-09-07T07:28:19.0332074Z * [new tag] ciflow/inductor/160706 -> ciflow/inductor/160706 2025-09-07T07:28:19.0332731Z * [new tag] ciflow/inductor/160729 -> ciflow/inductor/160729 2025-09-07T07:28:19.0333265Z * [new tag] ciflow/inductor/160798 -> ciflow/inductor/160798 2025-09-07T07:28:19.0333924Z * [new tag] ciflow/inductor/160836 -> ciflow/inductor/160836 2025-09-07T07:28:19.0334521Z * [new tag] ciflow/inductor/160843 -> ciflow/inductor/160843 2025-09-07T07:28:19.0335361Z * [new tag] ciflow/inductor/160869 -> ciflow/inductor/160869 2025-09-07T07:28:19.0335909Z * [new tag] ciflow/inductor/160920 -> ciflow/inductor/160920 2025-09-07T07:28:19.0336420Z * [new tag] ciflow/inductor/160943 -> ciflow/inductor/160943 2025-09-07T07:28:19.0336972Z * [new tag] ciflow/inductor/161092 -> ciflow/inductor/161092 2025-09-07T07:28:19.0337594Z * [new tag] ciflow/inductor/161093 -> ciflow/inductor/161093 2025-09-07T07:28:19.0338314Z * [new tag] ciflow/inductor/161109 -> ciflow/inductor/161109 2025-09-07T07:28:19.0338849Z * [new tag] ciflow/inductor/161118 -> ciflow/inductor/161118 2025-09-07T07:28:19.0339596Z * [new tag] ciflow/inductor/161178 -> ciflow/inductor/161178 2025-09-07T07:28:19.0340419Z * [new tag] ciflow/inductor/161246 -> ciflow/inductor/161246 2025-09-07T07:28:19.0340993Z * [new tag] ciflow/inductor/161349 -> ciflow/inductor/161349 2025-09-07T07:28:19.0341627Z * [new tag] ciflow/inductor/161350 -> ciflow/inductor/161350 2025-09-07T07:28:19.0342173Z * [new tag] ciflow/inductor/161351 -> ciflow/inductor/161351 2025-09-07T07:28:19.0342879Z * [new tag] ciflow/inductor/161397 -> ciflow/inductor/161397 2025-09-07T07:28:19.0343423Z * [new tag] ciflow/inductor/161404 -> ciflow/inductor/161404 2025-09-07T07:28:19.0344003Z * [new tag] ciflow/inductor/161405 -> ciflow/inductor/161405 2025-09-07T07:28:19.0344598Z * [new tag] ciflow/inductor/161406 -> ciflow/inductor/161406 2025-09-07T07:28:19.0345348Z * [new tag] ciflow/inductor/161410 -> ciflow/inductor/161410 2025-09-07T07:28:19.0345943Z * [new tag] ciflow/inductor/161414 -> ciflow/inductor/161414 2025-09-07T07:28:19.0346766Z * [new tag] ciflow/inductor/161442 -> ciflow/inductor/161442 2025-09-07T07:28:19.0347449Z * [new tag] ciflow/inductor/161458 -> ciflow/inductor/161458 2025-09-07T07:28:19.0347986Z * [new tag] ciflow/inductor/161468 -> ciflow/inductor/161468 2025-09-07T07:28:19.0348555Z * [new tag] ciflow/inductor/161469 -> ciflow/inductor/161469 2025-09-07T07:28:19.0349210Z * [new tag] ciflow/inductor/161485 -> ciflow/inductor/161485 2025-09-07T07:28:19.0349802Z * [new tag] ciflow/inductor/161499 -> ciflow/inductor/161499 2025-09-07T07:28:19.0350311Z * [new tag] ciflow/inductor/161534 -> ciflow/inductor/161534 2025-09-07T07:28:19.0350889Z * [new tag] ciflow/inductor/161595 -> ciflow/inductor/161595 2025-09-07T07:28:19.0351527Z * [new tag] ciflow/inductor/161596 -> ciflow/inductor/161596 2025-09-07T07:28:19.0352492Z * [new tag] ciflow/inductor/161630 -> ciflow/inductor/161630 2025-09-07T07:28:19.0353025Z * [new tag] ciflow/inductor/161667 -> ciflow/inductor/161667 2025-09-07T07:28:19.0353609Z * [new tag] ciflow/inductor/161670 -> ciflow/inductor/161670 2025-09-07T07:28:19.0354142Z * [new tag] ciflow/inductor/161673 -> ciflow/inductor/161673 2025-09-07T07:28:19.0354690Z * [new tag] ciflow/inductor/161674 -> ciflow/inductor/161674 2025-09-07T07:28:19.0355403Z * [new tag] ciflow/inductor/161675 -> ciflow/inductor/161675 2025-09-07T07:28:19.0355882Z * [new tag] ciflow/inductor/161693 -> ciflow/inductor/161693 2025-09-07T07:28:19.0356461Z * [new tag] ciflow/inductor/161695 -> ciflow/inductor/161695 2025-09-07T07:28:19.0357021Z * [new tag] ciflow/inductor/161715 -> ciflow/inductor/161715 2025-09-07T07:28:19.0357591Z * [new tag] ciflow/inductor/161730 -> ciflow/inductor/161730 2025-09-07T07:28:19.0358147Z * [new tag] ciflow/inductor/161732 -> ciflow/inductor/161732 2025-09-07T07:28:19.0358786Z * [new tag] ciflow/inductor/161744 -> ciflow/inductor/161744 2025-09-07T07:28:19.0359389Z * [new tag] ciflow/inductor/161746 -> ciflow/inductor/161746 2025-09-07T07:28:19.0359908Z * [new tag] ciflow/inductor/161747 -> ciflow/inductor/161747 2025-09-07T07:28:19.0360525Z * [new tag] ciflow/inductor/161819 -> ciflow/inductor/161819 2025-09-07T07:28:19.0361133Z * [new tag] ciflow/inductor/161821 -> ciflow/inductor/161821 2025-09-07T07:28:19.0361676Z * [new tag] ciflow/inductor/161828 -> ciflow/inductor/161828 2025-09-07T07:28:19.0371969Z * [new tag] ciflow/inductor/161879 -> ciflow/inductor/161879 2025-09-07T07:28:19.0372259Z * [new tag] ciflow/inductor/161880 -> ciflow/inductor/161880 2025-09-07T07:28:19.0372429Z * [new tag] ciflow/inductor/161881 -> ciflow/inductor/161881 2025-09-07T07:28:19.0372590Z * [new tag] ciflow/inductor/161907 -> ciflow/inductor/161907 2025-09-07T07:28:19.0372766Z * [new tag] ciflow/inductor/161914 -> ciflow/inductor/161914 2025-09-07T07:28:19.0372928Z * [new tag] ciflow/inductor/161924 -> ciflow/inductor/161924 2025-09-07T07:28:19.0373091Z * [new tag] ciflow/inductor/161936 -> ciflow/inductor/161936 2025-09-07T07:28:19.0373253Z * [new tag] ciflow/inductor/161938 -> ciflow/inductor/161938 2025-09-07T07:28:19.0373410Z * [new tag] ciflow/inductor/161939 -> ciflow/inductor/161939 2025-09-07T07:28:19.0373569Z * [new tag] ciflow/inductor/161940 -> ciflow/inductor/161940 2025-09-07T07:28:19.0373725Z * [new tag] ciflow/inductor/161955 -> ciflow/inductor/161955 2025-09-07T07:28:19.0373881Z * [new tag] ciflow/inductor/161957 -> ciflow/inductor/161957 2025-09-07T07:28:19.0374048Z * [new tag] ciflow/inductor/161975 -> ciflow/inductor/161975 2025-09-07T07:28:19.0374437Z * [new tag] ciflow/inductor/161977 -> ciflow/inductor/161977 2025-09-07T07:28:19.0374599Z * [new tag] ciflow/inductor/161978 -> ciflow/inductor/161978 2025-09-07T07:28:19.0374756Z * [new tag] ciflow/inductor/161979 -> ciflow/inductor/161979 2025-09-07T07:28:19.0374925Z * [new tag] ciflow/inductor/161980 -> ciflow/inductor/161980 2025-09-07T07:28:19.0375154Z * [new tag] ciflow/inductor/161988 -> ciflow/inductor/161988 2025-09-07T07:28:19.0375729Z * [new tag] ciflow/inductor/161994 -> ciflow/inductor/161994 2025-09-07T07:28:19.0376243Z * [new tag] ciflow/inductor/162013 -> ciflow/inductor/162013 2025-09-07T07:28:19.0376834Z * [new tag] ciflow/inductor/162014 -> ciflow/inductor/162014 2025-09-07T07:28:19.0377355Z * [new tag] ciflow/inductor/162017 -> ciflow/inductor/162017 2025-09-07T07:28:19.0377928Z * [new tag] ciflow/inductor/162021 -> ciflow/inductor/162021 2025-09-07T07:28:19.0378482Z * [new tag] ciflow/inductor/162023 -> ciflow/inductor/162023 2025-09-07T07:28:19.0379036Z * [new tag] ciflow/inductor/162027 -> ciflow/inductor/162027 2025-09-07T07:28:19.0379744Z * [new tag] ciflow/inductor/162029 -> ciflow/inductor/162029 2025-09-07T07:28:19.0380122Z * [new tag] ciflow/inductor/162030 -> ciflow/inductor/162030 2025-09-07T07:28:19.0380699Z * [new tag] ciflow/inductor/162031 -> ciflow/inductor/162031 2025-09-07T07:28:19.0381462Z * [new tag] ciflow/inductor/162033 -> ciflow/inductor/162033 2025-09-07T07:28:19.0382251Z * [new tag] ciflow/inductor/162052 -> ciflow/inductor/162052 2025-09-07T07:28:19.0382871Z * [new tag] ciflow/inductor/162053 -> ciflow/inductor/162053 2025-09-07T07:28:19.0383475Z * [new tag] ciflow/inductor/162056 -> ciflow/inductor/162056 2025-09-07T07:28:19.0384043Z * [new tag] ciflow/inductor/162063 -> ciflow/inductor/162063 2025-09-07T07:28:19.0384606Z * [new tag] ciflow/inductor/162066 -> ciflow/inductor/162066 2025-09-07T07:28:19.0385135Z * [new tag] ciflow/inductor/162068 -> ciflow/inductor/162068 2025-09-07T07:28:19.0386041Z * [new tag] ciflow/inductor/162081 -> ciflow/inductor/162081 2025-09-07T07:28:19.0386590Z * [new tag] ciflow/inductor/162088 -> ciflow/inductor/162088 2025-09-07T07:28:19.0387176Z * [new tag] ciflow/inductor/162089 -> ciflow/inductor/162089 2025-09-07T07:28:19.0387720Z * [new tag] ciflow/inductor/162094 -> ciflow/inductor/162094 2025-09-07T07:28:19.0388265Z * [new tag] ciflow/inductor/162098 -> ciflow/inductor/162098 2025-09-07T07:28:19.0388861Z * [new tag] ciflow/inductor/162101 -> ciflow/inductor/162101 2025-09-07T07:28:19.0389442Z * [new tag] ciflow/inductor/162102 -> ciflow/inductor/162102 2025-09-07T07:28:19.0389982Z * [new tag] ciflow/inductor/162104 -> ciflow/inductor/162104 2025-09-07T07:28:19.0390531Z * [new tag] ciflow/inductor/162106 -> ciflow/inductor/162106 2025-09-07T07:28:19.0391121Z * [new tag] ciflow/inductor/162108 -> ciflow/inductor/162108 2025-09-07T07:28:19.0391696Z * [new tag] ciflow/inductor/162126 -> ciflow/inductor/162126 2025-09-07T07:28:19.0392321Z * [new tag] ciflow/inductor/162149 -> ciflow/inductor/162149 2025-09-07T07:28:19.0392865Z * [new tag] ciflow/inductor/162164 -> ciflow/inductor/162164 2025-09-07T07:28:19.0393442Z * [new tag] ciflow/inductor/162166 -> ciflow/inductor/162166 2025-09-07T07:28:19.0393999Z * [new tag] ciflow/inductor/162169 -> ciflow/inductor/162169 2025-09-07T07:28:19.0394555Z * [new tag] ciflow/inductor/162170 -> ciflow/inductor/162170 2025-09-07T07:28:19.0395094Z * [new tag] ciflow/inductor/162171 -> ciflow/inductor/162171 2025-09-07T07:28:19.0395650Z * [new tag] ciflow/inductor/162183 -> ciflow/inductor/162183 2025-09-07T07:28:19.0396252Z * [new tag] ciflow/inductor/162189 -> ciflow/inductor/162189 2025-09-07T07:28:19.0396765Z * [new tag] ciflow/inductor/162190 -> ciflow/inductor/162190 2025-09-07T07:28:19.0397353Z * [new tag] ciflow/inductor/162191 -> ciflow/inductor/162191 2025-09-07T07:28:19.0397920Z * [new tag] ciflow/inductor/162194 -> ciflow/inductor/162194 2025-09-07T07:28:19.0398649Z * [new tag] ciflow/inductor/162200 -> ciflow/inductor/162200 2025-09-07T07:28:19.0399207Z * [new tag] ciflow/inductor/162201 -> ciflow/inductor/162201 2025-09-07T07:28:19.0399769Z * [new tag] ciflow/inductor/162208 -> ciflow/inductor/162208 2025-09-07T07:28:19.0400483Z * [new tag] ciflow/inductor/162211 -> ciflow/inductor/162211 2025-09-07T07:28:19.0401087Z * [new tag] ciflow/inductor/162216 -> ciflow/inductor/162216 2025-09-07T07:28:19.0401662Z * [new tag] ciflow/inductor/162220 -> ciflow/inductor/162220 2025-09-07T07:28:19.0402401Z * [new tag] ciflow/inductor/162222 -> ciflow/inductor/162222 2025-09-07T07:28:19.0402919Z * [new tag] ciflow/inductor/162227 -> ciflow/inductor/162227 2025-09-07T07:28:19.0403531Z * [new tag] ciflow/inductor/162238 -> ciflow/inductor/162238 2025-09-07T07:28:19.0404057Z * [new tag] ciflow/inductor/162239 -> ciflow/inductor/162239 2025-09-07T07:28:19.0404634Z * [new tag] ciflow/inductor/162240 -> ciflow/inductor/162240 2025-09-07T07:28:19.0405197Z * [new tag] ciflow/inductor/162244 -> ciflow/inductor/162244 2025-09-07T07:28:19.0405764Z * [new tag] ciflow/inductor/162245 -> ciflow/inductor/162245 2025-09-07T07:28:19.0406372Z * [new tag] ciflow/inductor/162262 -> ciflow/inductor/162262 2025-09-07T07:28:19.0406934Z * [new tag] ciflow/inductor/162275 -> ciflow/inductor/162275 2025-09-07T07:28:19.0407440Z * [new tag] ciflow/inductor/162278 -> ciflow/inductor/162278 2025-09-07T07:28:19.0408032Z * [new tag] ciflow/inductor/162284 -> ciflow/inductor/162284 2025-09-07T07:28:19.0408600Z * [new tag] ciflow/inductor/162286 -> ciflow/inductor/162286 2025-09-07T07:28:19.0409174Z * [new tag] ciflow/inductor/162288 -> ciflow/inductor/162288 2025-09-07T07:28:19.0409724Z * [new tag] ciflow/inductor/162293 -> ciflow/inductor/162293 2025-09-07T07:28:19.0410256Z * [new tag] ciflow/inductor/162294 -> ciflow/inductor/162294 2025-09-07T07:28:19.0410842Z * [new tag] ciflow/inductor/162295 -> ciflow/inductor/162295 2025-09-07T07:28:19.0411397Z * [new tag] ciflow/inductor/162296 -> ciflow/inductor/162296 2025-09-07T07:28:19.0411957Z * [new tag] ciflow/inductor/162298 -> ciflow/inductor/162298 2025-09-07T07:28:19.0412519Z * [new tag] ciflow/inductor/162307 -> ciflow/inductor/162307 2025-09-07T07:28:19.0413030Z * [new tag] ciflow/inductor/162309 -> ciflow/inductor/162309 2025-09-07T07:28:19.0413649Z * [new tag] ciflow/inductor/162311 -> ciflow/inductor/162311 2025-09-07T07:28:19.0414256Z * [new tag] ciflow/inductor/162312 -> ciflow/inductor/162312 2025-09-07T07:28:19.0414795Z * [new tag] ciflow/inductor/162315 -> ciflow/inductor/162315 2025-09-07T07:28:19.0415354Z * [new tag] ciflow/inductor/162316 -> ciflow/inductor/162316 2025-09-07T07:28:19.0415887Z * [new tag] ciflow/inductor/162318 -> ciflow/inductor/162318 2025-09-07T07:28:19.0416449Z * [new tag] ciflow/inductor/162323 -> ciflow/inductor/162323 2025-09-07T07:28:19.0417028Z * [new tag] ciflow/inductor/162341 -> ciflow/inductor/162341 2025-09-07T07:28:19.0417571Z * [new tag] ciflow/inductor/162345 -> ciflow/inductor/162345 2025-09-07T07:28:19.0418391Z * [new tag] ciflow/inductor/3b9a386 -> ciflow/inductor/3b9a386 2025-09-07T07:28:19.0419353Z * [new tag] ciflow/inductor/3d4b92b -> ciflow/inductor/3d4b92b 2025-09-07T07:28:19.0420056Z * [new tag] ciflow/inductor/d224ac7 -> ciflow/inductor/d224ac7 2025-09-07T07:28:19.0420920Z * [new tag] ciflow/linux-aarch64/157994 -> ciflow/linux-aarch64/157994 2025-09-07T07:28:19.0421497Z * [new tag] ciflow/linux-aarch64/159737 -> ciflow/linux-aarch64/159737 2025-09-07T07:28:19.0422014Z * [new tag] ciflow/linux-aarch64/160078 -> ciflow/linux-aarch64/160078 2025-09-07T07:28:19.0422699Z * [new tag] ciflow/mps/157553 -> ciflow/mps/157553 2025-09-07T07:28:19.0423290Z * [new tag] ciflow/mps/157635 -> ciflow/mps/157635 2025-09-07T07:28:19.0423689Z * [new tag] ciflow/mps/161988 -> ciflow/mps/161988 2025-09-07T07:28:19.0424239Z * [new tag] ciflow/mps/162108 -> ciflow/mps/162108 2025-09-07T07:28:19.0424736Z * [new tag] ciflow/mps/162153 -> ciflow/mps/162153 2025-09-07T07:28:19.0425281Z * [new tag] ciflow/mps/162281 -> ciflow/mps/162281 2025-09-07T07:28:19.0425944Z * [new tag] ciflow/nightly/156049 -> ciflow/nightly/156049 2025-09-07T07:28:19.0426458Z * [new tag] ciflow/nightly/158104 -> ciflow/nightly/158104 2025-09-07T07:28:19.0427154Z * [new tag] ciflow/op-benchmark/157994 -> ciflow/op-benchmark/157994 2025-09-07T07:28:19.0428018Z * [new tag] ciflow/periodic-rocm-mi300/161529 -> ciflow/periodic-rocm-mi300/161529 2025-09-07T07:28:19.0428593Z * [new tag] ciflow/periodic-rocm-mi300/161715 -> ciflow/periodic-rocm-mi300/161715 2025-09-07T07:28:19.0429423Z * [new tag] ciflow/periodic/054a2fd -> ciflow/periodic/054a2fd 2025-09-07T07:28:19.0430117Z * [new tag] ciflow/periodic/156703 -> ciflow/periodic/156703 2025-09-07T07:28:19.0430696Z * [new tag] ciflow/periodic/161715 -> ciflow/periodic/161715 2025-09-07T07:28:19.0431209Z * [new tag] ciflow/periodic/162021 -> ciflow/periodic/162021 2025-09-07T07:28:19.0431736Z * [new tag] ciflow/periodic/162323 -> ciflow/periodic/162323 2025-09-07T07:28:19.0432496Z * [new tag] ciflow/periodic/2a6d37d -> ciflow/periodic/2a6d37d 2025-09-07T07:28:19.0433142Z * [new tag] ciflow/periodic/317eeb8 -> ciflow/periodic/317eeb8 2025-09-07T07:28:19.0433741Z * [new tag] ciflow/periodic/3c32 -> ciflow/periodic/3c32 2025-09-07T07:28:19.0434505Z * [new tag] ciflow/periodic/3e98831 -> ciflow/periodic/3e98831 2025-09-07T07:28:19.0435258Z * [new tag] ciflow/periodic/94512-point -> ciflow/periodic/94512-point 2025-09-07T07:28:19.0436377Z * [new tag] ciflow/periodic/csl/test87519 -> ciflow/periodic/csl/test87519 2025-09-07T07:28:19.0437319Z * [new tag] ciflow/periodic/csltest88275 -> ciflow/periodic/csltest88275 2025-09-07T07:28:19.0438193Z * [new tag] ciflow/periodic/csltest88761 -> ciflow/periodic/csltest88761 2025-09-07T07:28:19.0439010Z * [new tag] ciflow/periodic/release_1.12 -> ciflow/periodic/release_1.12 2025-09-07T07:28:19.0439889Z * [new tag] ciflow/periodic/release_1.12.0 -> ciflow/periodic/release_1.12.0 2025-09-07T07:28:19.0440725Z * [new tag] ciflow/periodic/sha-ec5b83 -> ciflow/periodic/sha-ec5b83 2025-09-07T07:28:19.0441391Z * [new tag] ciflow/rocm-mi300/154170 -> ciflow/rocm-mi300/154170 2025-09-07T07:28:19.0442089Z * [new tag] ciflow/rocm-mi300/158747 -> ciflow/rocm-mi300/158747 2025-09-07T07:28:19.0442606Z * [new tag] ciflow/rocm-mi300/159146 -> ciflow/rocm-mi300/159146 2025-09-07T07:28:19.0443133Z * [new tag] ciflow/rocm-mi300/159158 -> ciflow/rocm-mi300/159158 2025-09-07T07:28:19.0443664Z * [new tag] ciflow/rocm-mi300/161715 -> ciflow/rocm-mi300/161715 2025-09-07T07:28:19.0444152Z * [new tag] ciflow/rocm-mi300/161957 -> ciflow/rocm-mi300/161957 2025-09-07T07:28:19.0444676Z * [new tag] ciflow/rocm-mi300/162053 -> ciflow/rocm-mi300/162053 2025-09-07T07:28:19.0445262Z * [new tag] ciflow/rocm-mi300/162056 -> ciflow/rocm-mi300/162056 2025-09-07T07:28:19.0445890Z * [new tag] ciflow/rocm-mi300/162112 -> ciflow/rocm-mi300/162112 2025-09-07T07:28:19.0446389Z * [new tag] ciflow/rocm-mi300/162245 -> ciflow/rocm-mi300/162245 2025-09-07T07:28:19.0447008Z * [new tag] ciflow/rocm-mi300/162278 -> ciflow/rocm-mi300/162278 2025-09-07T07:28:19.0447697Z * [new tag] ciflow/rocm-mi300/162288 -> ciflow/rocm-mi300/162288 2025-09-07T07:28:19.0448402Z * [new tag] ciflow/rocm-mi355/162053 -> ciflow/rocm-mi355/162053 2025-09-07T07:28:19.0448925Z * [new tag] ciflow/rocm-mi355/162056 -> ciflow/rocm-mi355/162056 2025-09-07T07:28:19.0449548Z * [new tag] ciflow/rocm/148492 -> ciflow/rocm/148492 2025-09-07T07:28:19.0450072Z * [new tag] ciflow/rocm/154170 -> ciflow/rocm/154170 2025-09-07T07:28:19.0450834Z * [new tag] ciflow/rocm/156491 -> ciflow/rocm/156491 2025-09-07T07:28:19.0451366Z * [new tag] ciflow/rocm/156592 -> ciflow/rocm/156592 2025-09-07T07:28:19.0451831Z * [new tag] ciflow/rocm/158747 -> ciflow/rocm/158747 2025-09-07T07:28:19.0452396Z * [new tag] ciflow/rocm/159146 -> ciflow/rocm/159146 2025-09-07T07:28:19.0453117Z * [new tag] ciflow/rocm/159158 -> ciflow/rocm/159158 2025-09-07T07:28:19.0453680Z * [new tag] ciflow/rocm/161715 -> ciflow/rocm/161715 2025-09-07T07:28:19.0454264Z * [new tag] ciflow/rocm/161972 -> ciflow/rocm/161972 2025-09-07T07:28:19.0454850Z * [new tag] ciflow/rocm/162052 -> ciflow/rocm/162052 2025-09-07T07:28:19.0455653Z * [new tag] ciflow/rocm/162053 -> ciflow/rocm/162053 2025-09-07T07:28:19.0456337Z * [new tag] ciflow/rocm/162056 -> ciflow/rocm/162056 2025-09-07T07:28:19.0457099Z * [new tag] ciflow/rocm/162112 -> ciflow/rocm/162112 2025-09-07T07:28:19.0457778Z * [new tag] ciflow/rocm/162278 -> ciflow/rocm/162278 2025-09-07T07:28:19.0458333Z * [new tag] ciflow/rocm/162288 -> ciflow/rocm/162288 2025-09-07T07:28:19.0458930Z * [new tag] ciflow/rocm/162305 -> ciflow/rocm/162305 2025-09-07T07:28:19.0459821Z * [new tag] ciflow/slow/01c7106 -> ciflow/slow/01c7106 2025-09-07T07:28:19.0460456Z * [new tag] ciflow/slow/0577043 -> ciflow/slow/0577043 2025-09-07T07:28:19.0461516Z * [new tag] ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym -> ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym 2025-09-07T07:28:19.0461909Z * [new tag] ciflow/slow/0e81104 -> ciflow/slow/0e81104 2025-09-07T07:28:19.0462477Z * [new tag] ciflow/slow/161395 -> ciflow/slow/161395 2025-09-07T07:28:19.0463185Z * [new tag] ciflow/slow/1732077 -> ciflow/slow/1732077 2025-09-07T07:28:19.0463933Z * [new tag] ciflow/slow/187eb7c -> ciflow/slow/187eb7c 2025-09-07T07:28:19.0464565Z * [new tag] ciflow/slow/1faef89 -> ciflow/slow/1faef89 2025-09-07T07:28:19.0465477Z * [new tag] ciflow/slow/3920ec1 -> ciflow/slow/3920ec1 2025-09-07T07:28:19.0466376Z * [new tag] ciflow/slow/3b7c6b2 -> ciflow/slow/3b7c6b2 2025-09-07T07:28:19.0467087Z * [new tag] ciflow/slow/59a3759 -> ciflow/slow/59a3759 2025-09-07T07:28:19.0467804Z * [new tag] ciflow/slow/70ef0bb -> ciflow/slow/70ef0bb 2025-09-07T07:28:19.0468469Z * [new tag] ciflow/slow/788ff06 -> ciflow/slow/788ff06 2025-09-07T07:28:19.0469594Z * [new tag] ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym -> ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym 2025-09-07T07:28:19.0470004Z * [new tag] ciflow/slow/9d85864 -> ciflow/slow/9d85864 2025-09-07T07:28:19.0470672Z * [new tag] ciflow/slow/9ffad5b -> ciflow/slow/9ffad5b 2025-09-07T07:28:19.0471569Z * [new tag] ciflow/slow/a206e8b -> ciflow/slow/a206e8b 2025-09-07T07:28:19.0472134Z * [new tag] ciflow/slow/a837609 -> ciflow/slow/a837609 2025-09-07T07:28:19.0472876Z * [new tag] ciflow/slow/af841f3 -> ciflow/slow/af841f3 2025-09-07T07:28:19.0473918Z * [new tag] ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym -> ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym 2025-09-07T07:28:19.0474443Z * [new tag] ciflow/triton_binaries/162329 -> ciflow/triton_binaries/162329 2025-09-07T07:28:19.0475068Z * [new tag] ciflow/trunk/113258 -> ciflow/trunk/113258 2025-09-07T07:28:19.0475590Z * [new tag] ciflow/trunk/137400 -> ciflow/trunk/137400 2025-09-07T07:28:19.0476174Z * [new tag] ciflow/trunk/148180 -> ciflow/trunk/148180 2025-09-07T07:28:19.0476657Z * [new tag] ciflow/trunk/148328 -> ciflow/trunk/148328 2025-09-07T07:28:19.0477182Z * [new tag] ciflow/trunk/148492 -> ciflow/trunk/148492 2025-09-07T07:28:19.0477915Z * [new tag] ciflow/trunk/148919 -> ciflow/trunk/148919 2025-09-07T07:28:19.0478484Z * [new tag] ciflow/trunk/152624 -> ciflow/trunk/152624 2025-09-07T07:28:19.0478964Z * [new tag] ciflow/trunk/154170 -> ciflow/trunk/154170 2025-09-07T07:28:19.0479502Z * [new tag] ciflow/trunk/154694 -> ciflow/trunk/154694 2025-09-07T07:28:19.0480059Z * [new tag] ciflow/trunk/156049 -> ciflow/trunk/156049 2025-09-07T07:28:19.0480576Z * [new tag] ciflow/trunk/156703 -> ciflow/trunk/156703 2025-09-07T07:28:19.0481310Z * [new tag] ciflow/trunk/156711 -> ciflow/trunk/156711 2025-09-07T07:28:19.0482030Z * [new tag] ciflow/trunk/157432 -> ciflow/trunk/157432 2025-09-07T07:28:19.0482647Z * [new tag] ciflow/trunk/157685 -> ciflow/trunk/157685 2025-09-07T07:28:19.0483212Z * [new tag] ciflow/trunk/157689 -> ciflow/trunk/157689 2025-09-07T07:28:19.0483802Z * [new tag] ciflow/trunk/157699 -> ciflow/trunk/157699 2025-09-07T07:28:19.0484341Z * [new tag] ciflow/trunk/157813 -> ciflow/trunk/157813 2025-09-07T07:28:19.0484889Z * [new tag] ciflow/trunk/157994 -> ciflow/trunk/157994 2025-09-07T07:28:19.0485439Z * [new tag] ciflow/trunk/158091 -> ciflow/trunk/158091 2025-09-07T07:28:19.0485977Z * [new tag] ciflow/trunk/158104 -> ciflow/trunk/158104 2025-09-07T07:28:19.0486527Z * [new tag] ciflow/trunk/158404 -> ciflow/trunk/158404 2025-09-07T07:28:19.0487150Z * [new tag] ciflow/trunk/158647 -> ciflow/trunk/158647 2025-09-07T07:28:19.0487941Z * [new tag] ciflow/trunk/158846 -> ciflow/trunk/158846 2025-09-07T07:28:19.0488472Z * [new tag] ciflow/trunk/159158 -> ciflow/trunk/159158 2025-09-07T07:28:19.0489200Z * [new tag] ciflow/trunk/159682 -> ciflow/trunk/159682 2025-09-07T07:28:19.0489759Z * [new tag] ciflow/trunk/159835 -> ciflow/trunk/159835 2025-09-07T07:28:19.0490320Z * [new tag] ciflow/trunk/160161 -> ciflow/trunk/160161 2025-09-07T07:28:19.0490882Z * [new tag] ciflow/trunk/160236 -> ciflow/trunk/160236 2025-09-07T07:28:19.0491401Z * [new tag] ciflow/trunk/160329 -> ciflow/trunk/160329 2025-09-07T07:28:19.0491936Z * [new tag] ciflow/trunk/160480 -> ciflow/trunk/160480 2025-09-07T07:28:19.0492522Z * [new tag] ciflow/trunk/160532 -> ciflow/trunk/160532 2025-09-07T07:28:19.0493071Z * [new tag] ciflow/trunk/160836 -> ciflow/trunk/160836 2025-09-07T07:28:19.0493706Z * [new tag] ciflow/trunk/160843 -> ciflow/trunk/160843 2025-09-07T07:28:19.0494179Z * [new tag] ciflow/trunk/160869 -> ciflow/trunk/160869 2025-09-07T07:28:19.0495001Z * [new tag] ciflow/trunk/160940 -> ciflow/trunk/160940 2025-09-07T07:28:19.0495572Z * [new tag] ciflow/trunk/160943 -> ciflow/trunk/160943 2025-09-07T07:28:19.0496317Z * [new tag] ciflow/trunk/160953 -> ciflow/trunk/160953 2025-09-07T07:28:19.0497050Z * [new tag] ciflow/trunk/161035 -> ciflow/trunk/161035 2025-09-07T07:28:19.0497587Z * [new tag] ciflow/trunk/161178 -> ciflow/trunk/161178 2025-09-07T07:28:19.0498141Z * [new tag] ciflow/trunk/161349 -> ciflow/trunk/161349 2025-09-07T07:28:19.0498759Z * [new tag] ciflow/trunk/161350 -> ciflow/trunk/161350 2025-09-07T07:28:19.0499294Z * [new tag] ciflow/trunk/161351 -> ciflow/trunk/161351 2025-09-07T07:28:19.0499839Z * [new tag] ciflow/trunk/161395 -> ciflow/trunk/161395 2025-09-07T07:28:19.0500360Z * [new tag] ciflow/trunk/161405 -> ciflow/trunk/161405 2025-09-07T07:28:19.0500899Z * [new tag] ciflow/trunk/161406 -> ciflow/trunk/161406 2025-09-07T07:28:19.0501441Z * [new tag] ciflow/trunk/161410 -> ciflow/trunk/161410 2025-09-07T07:28:19.0502032Z * [new tag] ciflow/trunk/161468 -> ciflow/trunk/161468 2025-09-07T07:28:19.0502529Z * [new tag] ciflow/trunk/161499 -> ciflow/trunk/161499 2025-09-07T07:28:19.0503365Z * [new tag] ciflow/trunk/161527 -> ciflow/trunk/161527 2025-09-07T07:28:19.0503942Z * [new tag] ciflow/trunk/161534 -> ciflow/trunk/161534 2025-09-07T07:28:19.0504499Z * [new tag] ciflow/trunk/161591 -> ciflow/trunk/161591 2025-09-07T07:28:19.0505088Z * [new tag] ciflow/trunk/161595 -> ciflow/trunk/161595 2025-09-07T07:28:19.0505592Z * [new tag] ciflow/trunk/161596 -> ciflow/trunk/161596 2025-09-07T07:28:19.0506152Z * [new tag] ciflow/trunk/161633 -> ciflow/trunk/161633 2025-09-07T07:28:19.0506687Z * [new tag] ciflow/trunk/161634 -> ciflow/trunk/161634 2025-09-07T07:28:19.0507286Z * [new tag] ciflow/trunk/161635 -> ciflow/trunk/161635 2025-09-07T07:28:19.0507818Z * [new tag] ciflow/trunk/161667 -> ciflow/trunk/161667 2025-09-07T07:28:19.0508337Z * [new tag] ciflow/trunk/161670 -> ciflow/trunk/161670 2025-09-07T07:28:19.0508905Z * [new tag] ciflow/trunk/161692 -> ciflow/trunk/161692 2025-09-07T07:28:19.0509572Z * [new tag] ciflow/trunk/161693 -> ciflow/trunk/161693 2025-09-07T07:28:19.0510296Z * [new tag] ciflow/trunk/161695 -> ciflow/trunk/161695 2025-09-07T07:28:19.0510824Z * [new tag] ciflow/trunk/161730 -> ciflow/trunk/161730 2025-09-07T07:28:19.0511374Z * [new tag] ciflow/trunk/161744 -> ciflow/trunk/161744 2025-09-07T07:28:19.0511972Z * [new tag] ciflow/trunk/161749 -> ciflow/trunk/161749 2025-09-07T07:28:19.0512494Z * [new tag] ciflow/trunk/161881 -> ciflow/trunk/161881 2025-09-07T07:28:19.0513068Z * [new tag] ciflow/trunk/161924 -> ciflow/trunk/161924 2025-09-07T07:28:19.0513844Z * [new tag] ciflow/trunk/161926 -> ciflow/trunk/161926 2025-09-07T07:28:19.0514478Z * [new tag] ciflow/trunk/161936 -> ciflow/trunk/161936 2025-09-07T07:28:19.0515023Z * [new tag] ciflow/trunk/161952 -> ciflow/trunk/161952 2025-09-07T07:28:19.0515622Z * [new tag] ciflow/trunk/161955 -> ciflow/trunk/161955 2025-09-07T07:28:19.0516123Z * [new tag] ciflow/trunk/161957 -> ciflow/trunk/161957 2025-09-07T07:28:19.0516646Z * [new tag] ciflow/trunk/161959 -> ciflow/trunk/161959 2025-09-07T07:28:19.0517246Z * [new tag] ciflow/trunk/161977 -> ciflow/trunk/161977 2025-09-07T07:28:19.0517793Z * [new tag] ciflow/trunk/161988 -> ciflow/trunk/161988 2025-09-07T07:28:19.0518308Z * [new tag] ciflow/trunk/161994 -> ciflow/trunk/161994 2025-09-07T07:28:19.0519002Z * [new tag] ciflow/trunk/162007 -> ciflow/trunk/162007 2025-09-07T07:28:19.0519532Z * [new tag] ciflow/trunk/162013 -> ciflow/trunk/162013 2025-09-07T07:28:19.0520127Z * [new tag] ciflow/trunk/162017 -> ciflow/trunk/162017 2025-09-07T07:28:19.0520704Z * [new tag] ciflow/trunk/162021 -> ciflow/trunk/162021 2025-09-07T07:28:19.0521285Z * [new tag] ciflow/trunk/162022 -> ciflow/trunk/162022 2025-09-07T07:28:19.0521792Z * [new tag] ciflow/trunk/162040 -> ciflow/trunk/162040 2025-09-07T07:28:19.0522392Z * [new tag] ciflow/trunk/162041 -> ciflow/trunk/162041 2025-09-07T07:28:19.0523084Z * [new tag] ciflow/trunk/162062 -> ciflow/trunk/162062 2025-09-07T07:28:19.0523627Z * [new tag] ciflow/trunk/162066 -> ciflow/trunk/162066 2025-09-07T07:28:19.0524159Z * [new tag] ciflow/trunk/162089 -> ciflow/trunk/162089 2025-09-07T07:28:19.0524760Z * [new tag] ciflow/trunk/162099 -> ciflow/trunk/162099 2025-09-07T07:28:19.0525295Z * [new tag] ciflow/trunk/162104 -> ciflow/trunk/162104 2025-09-07T07:28:19.0525836Z * [new tag] ciflow/trunk/162106 -> ciflow/trunk/162106 2025-09-07T07:28:19.0526421Z * [new tag] ciflow/trunk/162112 -> ciflow/trunk/162112 2025-09-07T07:28:19.0526930Z * [new tag] ciflow/trunk/162119 -> ciflow/trunk/162119 2025-09-07T07:28:19.0527516Z * [new tag] ciflow/trunk/162142 -> ciflow/trunk/162142 2025-09-07T07:28:19.0528044Z * [new tag] ciflow/trunk/162169 -> ciflow/trunk/162169 2025-09-07T07:28:19.0528620Z * [new tag] ciflow/trunk/162183 -> ciflow/trunk/162183 2025-09-07T07:28:19.0529180Z * [new tag] ciflow/trunk/162190 -> ciflow/trunk/162190 2025-09-07T07:28:19.0529702Z * [new tag] ciflow/trunk/162194 -> ciflow/trunk/162194 2025-09-07T07:28:19.0530386Z * [new tag] ciflow/trunk/162200 -> ciflow/trunk/162200 2025-09-07T07:28:19.0530956Z * [new tag] ciflow/trunk/162206 -> ciflow/trunk/162206 2025-09-07T07:28:19.0531462Z * [new tag] ciflow/trunk/162208 -> ciflow/trunk/162208 2025-09-07T07:28:19.0532117Z * [new tag] ciflow/trunk/162222 -> ciflow/trunk/162222 2025-09-07T07:28:19.0532678Z * [new tag] ciflow/trunk/162238 -> ciflow/trunk/162238 2025-09-07T07:28:19.0533271Z * [new tag] ciflow/trunk/162244 -> ciflow/trunk/162244 2025-09-07T07:28:19.0533992Z * [new tag] ciflow/trunk/162267 -> ciflow/trunk/162267 2025-09-07T07:28:19.0534661Z * [new tag] ciflow/trunk/162269 -> ciflow/trunk/162269 2025-09-07T07:28:19.0535221Z * [new tag] ciflow/trunk/162278 -> ciflow/trunk/162278 2025-09-07T07:28:19.0535790Z * [new tag] ciflow/trunk/162286 -> ciflow/trunk/162286 2025-09-07T07:28:19.0536337Z * [new tag] ciflow/trunk/162288 -> ciflow/trunk/162288 2025-09-07T07:28:19.0536893Z * [new tag] ciflow/trunk/162293 -> ciflow/trunk/162293 2025-09-07T07:28:19.0537553Z * [new tag] ciflow/trunk/162310 -> ciflow/trunk/162310 2025-09-07T07:28:19.0538007Z * [new tag] ciflow/trunk/162311 -> ciflow/trunk/162311 2025-09-07T07:28:19.0538538Z * [new tag] ciflow/trunk/162315 -> ciflow/trunk/162315 2025-09-07T07:28:19.0539129Z * [new tag] ciflow/trunk/162325 -> ciflow/trunk/162325 2025-09-07T07:28:19.0539862Z * [new tag] ciflow/trunk/162328 -> ciflow/trunk/162328 2025-09-07T07:28:19.0540454Z * [new tag] ciflow/trunk/162329 -> ciflow/trunk/162329 2025-09-07T07:28:19.0541344Z * [new tag] ciflow/unstable/123 -> ciflow/unstable/123 2025-09-07T07:28:19.0542075Z * [new tag] ciflow/vllm/162292 -> ciflow/vllm/162292 2025-09-07T07:28:19.0542756Z * [new tag] ciflow/win-arm64/156049 -> ciflow/win-arm64/156049 2025-09-07T07:28:19.0543251Z * [new tag] ciflow/win-arm64/158104 -> ciflow/win-arm64/158104 2025-09-07T07:28:19.0543890Z * [new tag] ciflow/xpu/157699 -> ciflow/xpu/157699 2025-09-07T07:28:19.0544412Z * [new tag] ciflow/xpu/157994 -> ciflow/xpu/157994 2025-09-07T07:28:19.0545057Z * [new tag] ciflow/xpu/159459 -> ciflow/xpu/159459 2025-09-07T07:28:19.0545621Z * [new tag] ciflow/xpu/159718 -> ciflow/xpu/159718 2025-09-07T07:28:19.0546159Z * [new tag] ciflow/xpu/159944 -> ciflow/xpu/159944 2025-09-07T07:28:19.0546787Z * [new tag] ciflow/xpu/160867 -> ciflow/xpu/160867 2025-09-07T07:28:19.0547420Z * [new tag] ciflow/xpu/160938 -> ciflow/xpu/160938 2025-09-07T07:28:19.0547902Z * [new tag] ciflow/xpu/160940 -> ciflow/xpu/160940 2025-09-07T07:28:19.0548467Z * [new tag] ciflow/xpu/160953 -> ciflow/xpu/160953 2025-09-07T07:28:19.0549161Z * [new tag] ciflow/xpu/161045 -> ciflow/xpu/161045 2025-09-07T07:28:19.0549844Z * [new tag] ciflow/xpu/161058 -> ciflow/xpu/161058 2025-09-07T07:28:19.0550625Z * [new tag] ciflow/xpu/161246 -> ciflow/xpu/161246 2025-09-07T07:28:19.0551352Z * [new tag] ciflow/xpu/161397 -> ciflow/xpu/161397 2025-09-07T07:28:19.0551991Z * [new tag] ciflow/xpu/161485 -> ciflow/xpu/161485 2025-09-07T07:28:19.0552606Z * [new tag] ciflow/xpu/161988 -> ciflow/xpu/161988 2025-09-07T07:28:19.0553120Z * [new tag] ciflow/xpu/162062 -> ciflow/xpu/162062 2025-09-07T07:28:19.0553796Z * [new tag] cslpull75 -> cslpull75 2025-09-07T07:28:19.0554393Z * [new tag] cslpull76 -> cslpull76 2025-09-07T07:28:19.0555055Z * [new tag] cslpull77 -> cslpull77 2025-09-07T07:28:19.0555693Z * [new tag] cslpull78 -> cslpull78 2025-09-07T07:28:19.0556687Z * [new tag] cslpull79 -> cslpull79 2025-09-07T07:28:19.0557555Z * [new tag] cslpull80 -> cslpull80 2025-09-07T07:28:19.0558256Z * [new tag] cslpull81 -> cslpull81 2025-09-07T07:28:19.0558931Z * [new tag] cslpull82 -> cslpull82 2025-09-07T07:28:19.0559583Z * [new tag] cslpull83 -> cslpull83 2025-09-07T07:28:19.0560251Z * [new tag] cslpull84 -> cslpull84 2025-09-07T07:28:19.0560863Z * [new tag] cslpull85 -> cslpull85 2025-09-07T07:28:19.0561535Z * [new tag] cslpull86 -> cslpull86 2025-09-07T07:28:19.0562222Z * [new tag] cslpull87 -> cslpull87 2025-09-07T07:28:19.0563072Z * [new tag] cslpull88 -> cslpull88 2025-09-07T07:28:19.0563627Z * [new tag] cslpull89 -> cslpull89 2025-09-07T07:28:19.0564188Z * [new tag] cslpull90 -> cslpull90 2025-09-07T07:28:19.0565412Z * [new tag] cslpull91 -> cslpull91 2025-09-07T07:28:19.0566093Z * [new tag] cslpull92 -> cslpull92 2025-09-07T07:28:19.0566771Z * [new tag] flight_5 -> flight_5 2025-09-07T07:28:19.0567552Z * [new tag] flight_5.1 -> flight_5.1 2025-09-07T07:28:19.0568244Z * [new tag] flight_5.2 -> flight_5.2 2025-09-07T07:28:19.0568847Z * [new tag] flight_5.3 -> flight_5.3 2025-09-07T07:28:19.0569513Z * [new tag] forpull1 -> forpull1 2025-09-07T07:28:19.0570470Z * [new tag] malfet/tag-2ef5611 -> malfet/tag-2ef5611 2025-09-07T07:28:19.0571222Z * [new tag] malfet/tag-317b1a0 -> malfet/tag-317b1a0 2025-09-07T07:28:19.0571920Z * [new tag] malfet/tag-ec6f767 -> malfet/tag-ec6f767 2025-09-07T07:28:19.0572625Z * [new tag] nightly-binary -> nightly-binary 2025-09-07T07:28:19.0573174Z * [new tag] sqzhang_flight4_plus -> sqzhang_flight4_plus 2025-09-07T07:28:19.0573914Z * [new tag] sqzhang_flight_3 -> sqzhang_flight_3 2025-09-07T07:28:19.0574923Z * [new tag] trunk/00636e0171e7e733628c408084805442270cf608 -> trunk/00636e0171e7e733628c408084805442270cf608 2025-09-07T07:28:19.0575694Z * [new tag] trunk/019fed39aa6b2dd8c69347378d53423e5efae8d4 -> trunk/019fed39aa6b2dd8c69347378d53423e5efae8d4 2025-09-07T07:28:19.0576640Z * [new tag] trunk/01ab325cc2e0dc221af4d710974e1b9175066544 -> trunk/01ab325cc2e0dc221af4d710974e1b9175066544 2025-09-07T07:28:19.0577651Z * [new tag] trunk/01edcd4df8bf0c7b4cc2d3ec868bd2059eeea83b -> trunk/01edcd4df8bf0c7b4cc2d3ec868bd2059eeea83b 2025-09-07T07:28:19.0578413Z * [new tag] trunk/040d00af048967dde7938d358d7f5988cbd18388 -> trunk/040d00af048967dde7938d358d7f5988cbd18388 2025-09-07T07:28:19.0579176Z * [new tag] trunk/0447f2d99b4351b2ff129dce6eebb371024f73e5 -> trunk/0447f2d99b4351b2ff129dce6eebb371024f73e5 2025-09-07T07:28:19.0579866Z * [new tag] trunk/047603d35bdc70046216384838d6340feab79bf4 -> trunk/047603d35bdc70046216384838d6340feab79bf4 2025-09-07T07:28:19.0580626Z * [new tag] trunk/06da7c0730b3764f178ec3a90dedf4ffa4202d81 -> trunk/06da7c0730b3764f178ec3a90dedf4ffa4202d81 2025-09-07T07:28:19.0581520Z * [new tag] trunk/081cab045472ce045634548cc6c14a4870641e23 -> trunk/081cab045472ce045634548cc6c14a4870641e23 2025-09-07T07:28:19.0582206Z * [new tag] trunk/09587daf8c9f21f5340f73921ce5f23d1a4a4572 -> trunk/09587daf8c9f21f5340f73921ce5f23d1a4a4572 2025-09-07T07:28:19.0582839Z * [new tag] trunk/09be1890d72cc34fc946965dc4a27736bf0ca8c6 -> trunk/09be1890d72cc34fc946965dc4a27736bf0ca8c6 2025-09-07T07:28:19.0583564Z * [new tag] trunk/09d2f1b6315d6d416fbf452793d65795863ebc66 -> trunk/09d2f1b6315d6d416fbf452793d65795863ebc66 2025-09-07T07:28:19.0584265Z * [new tag] trunk/0af70e2353e1dcda83175fd4834ecb7b63e009e0 -> trunk/0af70e2353e1dcda83175fd4834ecb7b63e009e0 2025-09-07T07:28:19.0585501Z * [new tag] trunk/0c0e056a9e20c17271a6144dd32c0c7e3ba26736 -> trunk/0c0e056a9e20c17271a6144dd32c0c7e3ba26736 2025-09-07T07:28:19.0586224Z * [new tag] trunk/0cd6c56bdfa9178ff61be82ce3b178926ddb64a9 -> trunk/0cd6c56bdfa9178ff61be82ce3b178926ddb64a9 2025-09-07T07:28:19.0586928Z * [new tag] trunk/0d421ace32c1605ee8e452ee1eeb03bd243dd96c -> trunk/0d421ace32c1605ee8e452ee1eeb03bd243dd96c 2025-09-07T07:28:19.0587907Z * [new tag] trunk/0d71a9dd5b4b6d1dde58d91c9b71d96bc6a6a171 -> trunk/0d71a9dd5b4b6d1dde58d91c9b71d96bc6a6a171 2025-09-07T07:28:19.0588472Z * [new tag] trunk/0d84ff3b78f55492d3d4708458c92d776274939e -> trunk/0d84ff3b78f55492d3d4708458c92d776274939e 2025-09-07T07:28:19.0589165Z * [new tag] trunk/0f45aaf4414048b17d720d0915ce221a8de8ec63 -> trunk/0f45aaf4414048b17d720d0915ce221a8de8ec63 2025-09-07T07:28:19.0589857Z * [new tag] trunk/0ff8eabf1387de5acd6712a03bda61f1a3dfa27f -> trunk/0ff8eabf1387de5acd6712a03bda61f1a3dfa27f 2025-09-07T07:28:19.0590566Z * [new tag] trunk/104f2680e03d13a4765ca69f905d8f16fc0c822f -> trunk/104f2680e03d13a4765ca69f905d8f16fc0c822f 2025-09-07T07:28:19.0591289Z * [new tag] trunk/12814701555d3e41dfcdf8f9273af5821e322df0 -> trunk/12814701555d3e41dfcdf8f9273af5821e322df0 2025-09-07T07:28:19.0591967Z * [new tag] trunk/13b65196db422bdb394cb482e208c61ed448898c -> trunk/13b65196db422bdb394cb482e208c61ed448898c 2025-09-07T07:28:19.0592667Z * [new tag] trunk/13d66e2a66eceed14b8a8f5a971087df4f688a46 -> trunk/13d66e2a66eceed14b8a8f5a971087df4f688a46 2025-09-07T07:28:19.0593467Z * [new tag] trunk/145a3a7bda15e3963a33eb1b54bba5d4a270b225 -> trunk/145a3a7bda15e3963a33eb1b54bba5d4a270b225 2025-09-07T07:28:19.0594115Z * [new tag] trunk/146371483318e17929daefd37c8e459d9d6d47bb -> trunk/146371483318e17929daefd37c8e459d9d6d47bb 2025-09-07T07:28:19.0594921Z * [new tag] trunk/15c77a8cfd341e74fd124b077492ef2bfa51b339 -> trunk/15c77a8cfd341e74fd124b077492ef2bfa51b339 2025-09-07T07:28:19.0595610Z * [new tag] trunk/17fa8eec4a1e32939ab4d364ee6e75487a79b654 -> trunk/17fa8eec4a1e32939ab4d364ee6e75487a79b654 2025-09-07T07:28:19.0596792Z * [new tag] trunk/190c391a28845a14df26abb228d26aa813efb20c -> trunk/190c391a28845a14df26abb228d26aa813efb20c 2025-09-07T07:28:19.0597510Z * [new tag] trunk/1a588ace4667bde1331fbd8ed957157dca5cee68 -> trunk/1a588ace4667bde1331fbd8ed957157dca5cee68 2025-09-07T07:28:19.0598241Z * [new tag] trunk/1aa7476885e8f6e7b0ec3a5b6383aad9d3f343e7 -> trunk/1aa7476885e8f6e7b0ec3a5b6383aad9d3f343e7 2025-09-07T07:28:19.0598851Z * [new tag] trunk/1aeb421c342c9e9607842f4c87cb46e8e816ee53 -> trunk/1aeb421c342c9e9607842f4c87cb46e8e816ee53 2025-09-07T07:28:19.0599580Z * [new tag] trunk/1c1b28d5b6a942fafe23b2f09302d93c25226d4a -> trunk/1c1b28d5b6a942fafe23b2f09302d93c25226d4a 2025-09-07T07:28:19.0600363Z * [new tag] trunk/1ebd70d0c0d562d3be9abdee2a21906584af7d99 -> trunk/1ebd70d0c0d562d3be9abdee2a21906584af7d99 2025-09-07T07:28:19.0601098Z * [new tag] trunk/1ec2c15914da4ef7bd926ed9aebc8671c75fe965 -> trunk/1ec2c15914da4ef7bd926ed9aebc8671c75fe965 2025-09-07T07:28:19.0601746Z * [new tag] trunk/1f51056bd64e73d1aa81321bc3c098575b1bc78a -> trunk/1f51056bd64e73d1aa81321bc3c098575b1bc78a 2025-09-07T07:28:19.0602484Z * [new tag] trunk/1f820de639c75a1562d3fb03f160439f853ae07b -> trunk/1f820de639c75a1562d3fb03f160439f853ae07b 2025-09-07T07:28:19.0603177Z * [new tag] trunk/204697f0e695d82894c5010fbec664c4391f90cc -> trunk/204697f0e695d82894c5010fbec664c4391f90cc 2025-09-07T07:28:19.0603864Z * [new tag] trunk/20629b1619fe636227d01fc85ba221daa7185a05 -> trunk/20629b1619fe636227d01fc85ba221daa7185a05 2025-09-07T07:28:19.0604615Z * [new tag] trunk/20b47acef845e9c4f71da9429a396d293f50ebe7 -> trunk/20b47acef845e9c4f71da9429a396d293f50ebe7 2025-09-07T07:28:19.0605634Z * [new tag] trunk/20bfb2539d7c5250379648eda35f80b8a7d642dd -> trunk/20bfb2539d7c5250379648eda35f80b8a7d642dd 2025-09-07T07:28:19.0606402Z * [new tag] trunk/21fae99c180d17def562797ea0fb154d8fdf88e3 -> trunk/21fae99c180d17def562797ea0fb154d8fdf88e3 2025-09-07T07:28:19.0607215Z * [new tag] trunk/248355faf53f9f7ba2fd0a367d59600c6d991e7f -> trunk/248355faf53f9f7ba2fd0a367d59600c6d991e7f 2025-09-07T07:28:19.0607771Z * [new tag] trunk/25f4aaed9ec26f39c13862323ff8582006473d23 -> trunk/25f4aaed9ec26f39c13862323ff8582006473d23 2025-09-07T07:28:19.0608517Z * [new tag] trunk/261a84a1764412f8e659c956e3f81997ec3de9d5 -> trunk/261a84a1764412f8e659c956e3f81997ec3de9d5 2025-09-07T07:28:19.0609312Z * [new tag] trunk/28f4ab0737937858730f29f5c4e601e109cf9d5f -> trunk/28f4ab0737937858730f29f5c4e601e109cf9d5f 2025-09-07T07:28:19.0610080Z * [new tag] trunk/291cd11f2d5df6f48d348cce0e4e762f274f4dc4 -> trunk/291cd11f2d5df6f48d348cce0e4e762f274f4dc4 2025-09-07T07:28:19.0610775Z * [new tag] trunk/29280864d941e6108ab57f7298f520c0cf9696e9 -> trunk/29280864d941e6108ab57f7298f520c0cf9696e9 2025-09-07T07:28:19.0611587Z * [new tag] trunk/2a45837e98c63cae9d1a2e2133a727b829e549d5 -> trunk/2a45837e98c63cae9d1a2e2133a727b829e549d5 2025-09-07T07:28:19.0612359Z * [new tag] trunk/2a5c0785e2f975697fd7bdf1411de6e03dcaa1ef -> trunk/2a5c0785e2f975697fd7bdf1411de6e03dcaa1ef 2025-09-07T07:28:19.0613068Z * [new tag] trunk/2b8a83901c58a0858ea9e4ce00055f48e6ed164c -> trunk/2b8a83901c58a0858ea9e4ce00055f48e6ed164c 2025-09-07T07:28:19.0613707Z * [new tag] trunk/2ba65472dd54488a86a50326ea990195fc6732d6 -> trunk/2ba65472dd54488a86a50326ea990195fc6732d6 2025-09-07T07:28:19.0614394Z * [new tag] trunk/2c03f0acc53ed13fe8ebfe809129f25996e009a0 -> trunk/2c03f0acc53ed13fe8ebfe809129f25996e009a0 2025-09-07T07:28:19.0615121Z * [new tag] trunk/2dd529df0092799f68ee7afcf52338276906706a -> trunk/2dd529df0092799f68ee7afcf52338276906706a 2025-09-07T07:28:19.0616873Z * [new tag] trunk/2f6b4b1ad3f82bb3bd984f6e65744ea339ffb8b5 -> trunk/2f6b4b1ad3f82bb3bd984f6e65744ea339ffb8b5 2025-09-07T07:28:19.0617207Z * [new tag] trunk/2fa0520a64ed8aa734a56c4d124958f0b5711ca8 -> trunk/2fa0520a64ed8aa734a56c4d124958f0b5711ca8 2025-09-07T07:28:19.0617533Z * [new tag] trunk/302df2ac5dc4222294c09d48804a2dddb8f4bad8 -> trunk/302df2ac5dc4222294c09d48804a2dddb8f4bad8 2025-09-07T07:28:19.0618034Z * [new tag] trunk/33028597bfa2e0178e28c8cce33cb9b3800cac43 -> trunk/33028597bfa2e0178e28c8cce33cb9b3800cac43 2025-09-07T07:28:19.0618694Z * [new tag] trunk/34aa78274d6770086025a967fa63a86830e08176 -> trunk/34aa78274d6770086025a967fa63a86830e08176 2025-09-07T07:28:19.0619284Z * [new tag] trunk/3559c354ce6a14d11fe29fb12fa2747a2f2af449 -> trunk/3559c354ce6a14d11fe29fb12fa2747a2f2af449 2025-09-07T07:28:19.0619910Z * [new tag] trunk/36d207fcaaede0d1e58a5168084c307b32b6fd8b -> trunk/36d207fcaaede0d1e58a5168084c307b32b6fd8b 2025-09-07T07:28:19.0620488Z * [new tag] trunk/377033757ae5ca524ea842f1b0a5f446ed3d8fe0 -> trunk/377033757ae5ca524ea842f1b0a5f446ed3d8fe0 2025-09-07T07:28:19.0621148Z * [new tag] trunk/3771380f83fcac154a7c89ad679311d8c4818287 -> trunk/3771380f83fcac154a7c89ad679311d8c4818287 2025-09-07T07:28:19.0621917Z * [new tag] trunk/3a207816cc569f78863d86c01f2a3d265350e39f -> trunk/3a207816cc569f78863d86c01f2a3d265350e39f 2025-09-07T07:28:19.0622580Z * [new tag] trunk/3a20a20e7065ec927fdd216d4da3b04f879b3c67 -> trunk/3a20a20e7065ec927fdd216d4da3b04f879b3c67 2025-09-07T07:28:19.0623390Z * [new tag] trunk/3bbc2e3e4f025523eaa5dbff220b3e96bca608d0 -> trunk/3bbc2e3e4f025523eaa5dbff220b3e96bca608d0 2025-09-07T07:28:19.0624029Z * [new tag] trunk/3c0ff1b569c45cfa6935ad8031a9d4cf1551aa3f -> trunk/3c0ff1b569c45cfa6935ad8031a9d4cf1551aa3f 2025-09-07T07:28:19.0624984Z * [new tag] trunk/3c45af079afc92a03b03ddf4f9198902ffcf30cf -> trunk/3c45af079afc92a03b03ddf4f9198902ffcf30cf 2025-09-07T07:28:19.0625734Z * [new tag] trunk/3dde5d7f9bf80dd6623a712bc429e9e4302464b5 -> trunk/3dde5d7f9bf80dd6623a712bc429e9e4302464b5 2025-09-07T07:28:19.0626406Z * [new tag] trunk/403a3a393cda7e60f503f3b04b8805a845dcf45d -> trunk/403a3a393cda7e60f503f3b04b8805a845dcf45d 2025-09-07T07:28:19.0627167Z * [new tag] trunk/420c52ecf36f86d32da0853bfbe074b682b070aa -> trunk/420c52ecf36f86d32da0853bfbe074b682b070aa 2025-09-07T07:28:19.0627848Z * [new tag] trunk/43b7c86a2c0f91320f5c5f4827b111edff06fdb6 -> trunk/43b7c86a2c0f91320f5c5f4827b111edff06fdb6 2025-09-07T07:28:19.0628518Z * [new tag] trunk/451ed931562ec8b46d1f7e6c266a68132a119336 -> trunk/451ed931562ec8b46d1f7e6c266a68132a119336 2025-09-07T07:28:19.0629213Z * [new tag] trunk/480c7391126656154318fabf1d57ebc01e196e63 -> trunk/480c7391126656154318fabf1d57ebc01e196e63 2025-09-07T07:28:19.0629969Z * [new tag] trunk/48bedd753da22634aa94fbafeb731e82025404f3 -> trunk/48bedd753da22634aa94fbafeb731e82025404f3 2025-09-07T07:28:19.0630617Z * [new tag] trunk/494878a11b79071ada0b98f34042d47155be6d1c -> trunk/494878a11b79071ada0b98f34042d47155be6d1c 2025-09-07T07:28:19.0631408Z * [new tag] trunk/4ae57d448c0a7d37e4cfd5c27d977fad2cef4051 -> trunk/4ae57d448c0a7d37e4cfd5c27d977fad2cef4051 2025-09-07T07:28:19.0632148Z * [new tag] trunk/4cdaf8265d86f984254b62052da8c26ef61ef1cf -> trunk/4cdaf8265d86f984254b62052da8c26ef61ef1cf 2025-09-07T07:28:19.0632697Z * [new tag] trunk/4d4abec80f03cd8fdefe1d9cb3a60d3690cd777e -> trunk/4d4abec80f03cd8fdefe1d9cb3a60d3690cd777e 2025-09-07T07:28:19.0633438Z * [new tag] trunk/4e42aa8ffc44b8340eb0eeaf80a2cafc4763a186 -> trunk/4e42aa8ffc44b8340eb0eeaf80a2cafc4763a186 2025-09-07T07:28:19.0634138Z * [new tag] trunk/4f72d932feee0749397fec876dcd43994f50b215 -> trunk/4f72d932feee0749397fec876dcd43994f50b215 2025-09-07T07:28:19.0634913Z * [new tag] trunk/50fc22dedf3c4a27be61fa05551c4f320281b42d -> trunk/50fc22dedf3c4a27be61fa05551c4f320281b42d 2025-09-07T07:28:19.0635614Z * [new tag] trunk/5211f1f908907ffc064b56e43cf8659f7fc22aa9 -> trunk/5211f1f908907ffc064b56e43cf8659f7fc22aa9 2025-09-07T07:28:19.0636354Z * [new tag] trunk/524b78d4f67045b83bb69edc56ab16efe282971c -> trunk/524b78d4f67045b83bb69edc56ab16efe282971c 2025-09-07T07:28:19.0637116Z * [new tag] trunk/54e275e0d81fe1e1ccfa4fb5f2a5a9aaca00ca15 -> trunk/54e275e0d81fe1e1ccfa4fb5f2a5a9aaca00ca15 2025-09-07T07:28:19.0637756Z * [new tag] trunk/5561e45758d59c94605873d5db48ed459c004c3b -> trunk/5561e45758d59c94605873d5db48ed459c004c3b 2025-09-07T07:28:19.0638608Z * [new tag] trunk/57278d45f046d4f89f45d373b1af4dd56934ff24 -> trunk/57278d45f046d4f89f45d373b1af4dd56934ff24 2025-09-07T07:28:19.0639316Z * [new tag] trunk/5927a70934ccf7b70182d364c23245a7dd685503 -> trunk/5927a70934ccf7b70182d364c23245a7dd685503 2025-09-07T07:28:19.0640073Z * [new tag] trunk/5985e28912aeb40b103ebfcf2fd0665eb4a50599 -> trunk/5985e28912aeb40b103ebfcf2fd0665eb4a50599 2025-09-07T07:28:19.0640789Z * [new tag] trunk/5a2da090ed6db88bb657c4e51ec0b310cd08bff6 -> trunk/5a2da090ed6db88bb657c4e51ec0b310cd08bff6 2025-09-07T07:28:19.0641559Z * [new tag] trunk/5c473e9f5ee0ef0fc38e6cf34a95b547f8cdc8d5 -> trunk/5c473e9f5ee0ef0fc38e6cf34a95b547f8cdc8d5 2025-09-07T07:28:19.0642226Z * [new tag] trunk/5c67426d6847667a7c55a2dd01f470fa37238c18 -> trunk/5c67426d6847667a7c55a2dd01f470fa37238c18 2025-09-07T07:28:19.0642916Z * [new tag] trunk/5da573c42c332bc68d4b7946c69f690a876d951a -> trunk/5da573c42c332bc68d4b7946c69f690a876d951a 2025-09-07T07:28:19.0643702Z * [new tag] trunk/5e5870e858f60ff4bf87d03f3592097e934a9580 -> trunk/5e5870e858f60ff4bf87d03f3592097e934a9580 2025-09-07T07:28:19.0644420Z * [new tag] trunk/5f3cbc9442aa55b5afb29f4ac8ca9be569003e84 -> trunk/5f3cbc9442aa55b5afb29f4ac8ca9be569003e84 2025-09-07T07:28:19.0645184Z * [new tag] trunk/600c25e9a17fe56e3dee872be8854db08916ba0c -> trunk/600c25e9a17fe56e3dee872be8854db08916ba0c 2025-09-07T07:28:19.0645836Z * [new tag] trunk/601ae8e4831fc8123fffcfb8fd2e6b6381b42e14 -> trunk/601ae8e4831fc8123fffcfb8fd2e6b6381b42e14 2025-09-07T07:28:19.0646527Z * [new tag] trunk/6087ef41e54c2494b117ffd923faf20f515a6806 -> trunk/6087ef41e54c2494b117ffd923faf20f515a6806 2025-09-07T07:28:19.0647279Z * [new tag] trunk/626cb7df8161dd4ecb4fe43b60f37ce9076f56b1 -> trunk/626cb7df8161dd4ecb4fe43b60f37ce9076f56b1 2025-09-07T07:28:19.0647927Z * [new tag] trunk/62c3f9a97fd3dea7132a93066d32d893ffe101e6 -> trunk/62c3f9a97fd3dea7132a93066d32d893ffe101e6 2025-09-07T07:28:19.0648690Z * [new tag] trunk/63a9c23fe99eacfd09610c36dfe8f01b053c1a35 -> trunk/63a9c23fe99eacfd09610c36dfe8f01b053c1a35 2025-09-07T07:28:19.0649432Z * [new tag] trunk/65985937d97505f648b6ed852c3129f2dd08b251 -> trunk/65985937d97505f648b6ed852c3129f2dd08b251 2025-09-07T07:28:19.0650634Z * [new tag] trunk/66f3b4a682a6153517dd23369fdc3289b6494b07 -> trunk/66f3b4a682a6153517dd23369fdc3289b6494b07 2025-09-07T07:28:19.0651183Z * [new tag] trunk/6737e2c996990024187ba620d2764f3b6f6add2c -> trunk/6737e2c996990024187ba620d2764f3b6f6add2c 2025-09-07T07:28:19.0651866Z * [new tag] trunk/67c31dcd364f10072a55f4a30ffd1151c686283a -> trunk/67c31dcd364f10072a55f4a30ffd1151c686283a 2025-09-07T07:28:19.0652593Z * [new tag] trunk/68738beff73e9c3512e18b4edea811a897ce42db -> trunk/68738beff73e9c3512e18b4edea811a897ce42db 2025-09-07T07:28:19.0653370Z * [new tag] trunk/69a25f68884a168550695fdb1a7c310c54d29536 -> trunk/69a25f68884a168550695fdb1a7c310c54d29536 2025-09-07T07:28:19.0654060Z * [new tag] trunk/6b1900c22f1a07b9519346898d4c71d8a2b0f12f -> trunk/6b1900c22f1a07b9519346898d4c71d8a2b0f12f 2025-09-07T07:28:19.0654762Z * [new tag] trunk/6b8b3ac4403f771bd4a8f9a45d93347304148774 -> trunk/6b8b3ac4403f771bd4a8f9a45d93347304148774 2025-09-07T07:28:19.0655507Z * [new tag] trunk/6f7608d603834d6068b2e7a5d59bec3973b6bb1b -> trunk/6f7608d603834d6068b2e7a5d59bec3973b6bb1b 2025-09-07T07:28:19.0656253Z * [new tag] trunk/70d36e047dfb3488fd6335016711a784d810ebda -> trunk/70d36e047dfb3488fd6335016711a784d810ebda 2025-09-07T07:28:19.0656926Z * [new tag] trunk/71992dd805ff9d6763f77214dfe8b0465e88c87b -> trunk/71992dd805ff9d6763f77214dfe8b0465e88c87b 2025-09-07T07:28:19.0657603Z * [new tag] trunk/734ce8eba9c69381f187359bf0fef1d71d84cd20 -> trunk/734ce8eba9c69381f187359bf0fef1d71d84cd20 2025-09-07T07:28:19.0658354Z * [new tag] trunk/73eb4511fb863a37944342b7e92aae706de603c8 -> trunk/73eb4511fb863a37944342b7e92aae706de603c8 2025-09-07T07:28:19.0659143Z * [new tag] trunk/75bc23cfc345bd4c05e7f97c416c4b3d2d1fa64b -> trunk/75bc23cfc345bd4c05e7f97c416c4b3d2d1fa64b 2025-09-07T07:28:19.0659806Z * [new tag] trunk/771f369448321a387f2018535bc8b8b6e5f12fab -> trunk/771f369448321a387f2018535bc8b8b6e5f12fab 2025-09-07T07:28:19.0660629Z * [new tag] trunk/789d4942127143f2adcb53612c058ce4c9a2cf20 -> trunk/789d4942127143f2adcb53612c058ce4c9a2cf20 2025-09-07T07:28:19.0661208Z * [new tag] trunk/791eff96c85678c950888f9da24650083ee673fe -> trunk/791eff96c85678c950888f9da24650083ee673fe 2025-09-07T07:28:19.0661792Z * [new tag] trunk/793fc12aff1f69fbbf9f4278182fb52bbe350fc9 -> trunk/793fc12aff1f69fbbf9f4278182fb52bbe350fc9 2025-09-07T07:28:19.0662493Z * [new tag] trunk/79fcd5247a9a129eee526a14df30bfc6a22b3f01 -> trunk/79fcd5247a9a129eee526a14df30bfc6a22b3f01 2025-09-07T07:28:19.0663146Z * [new tag] trunk/7f4ff79210eb06924f223ae3a1941ee0e2635348 -> trunk/7f4ff79210eb06924f223ae3a1941ee0e2635348 2025-09-07T07:28:19.0663963Z * [new tag] trunk/8076a185c85112be62be292eb47409c88a585b1c -> trunk/8076a185c85112be62be292eb47409c88a585b1c 2025-09-07T07:28:19.0664607Z * [new tag] trunk/80dd397f1979371a5583fa3d5c7352029522a78d -> trunk/80dd397f1979371a5583fa3d5c7352029522a78d 2025-09-07T07:28:19.0665197Z * [new tag] trunk/8171d6052ec12628eb67e0040839314056014429 -> trunk/8171d6052ec12628eb67e0040839314056014429 2025-09-07T07:28:19.0666011Z * [new tag] trunk/81aeefa657b7ccc26b275c50a9f33b2f056e8071 -> trunk/81aeefa657b7ccc26b275c50a9f33b2f056e8071 2025-09-07T07:28:19.0666744Z * [new tag] trunk/81b7b16618bda250ce55982894a83dc0805eb64c -> trunk/81b7b16618bda250ce55982894a83dc0805eb64c 2025-09-07T07:28:19.0667483Z * [new tag] trunk/827f0d405448de31f79d1089f7d7fceab2f87895 -> trunk/827f0d405448de31f79d1089f7d7fceab2f87895 2025-09-07T07:28:19.0668209Z * [new tag] trunk/82f63c8f6de63c30132a8ac299b6e8c2fd0d3fe8 -> trunk/82f63c8f6de63c30132a8ac299b6e8c2fd0d3fe8 2025-09-07T07:28:19.0668990Z * [new tag] trunk/850e1382a9c56bfde18af09d3e72352d775e9435 -> trunk/850e1382a9c56bfde18af09d3e72352d775e9435 2025-09-07T07:28:19.0669730Z * [new tag] trunk/8678d831c48e616b717bff50f2d03141d2e9f965 -> trunk/8678d831c48e616b717bff50f2d03141d2e9f965 2025-09-07T07:28:19.0670539Z * [new tag] trunk/869cbcc16e489a4f5a14a93d5779b0ea86061c60 -> trunk/869cbcc16e489a4f5a14a93d5779b0ea86061c60 2025-09-07T07:28:19.0671487Z * [new tag] trunk/8703debf669bc2238211bfd039f4ecdd8228b7f7 -> trunk/8703debf669bc2238211bfd039f4ecdd8228b7f7 2025-09-07T07:28:19.0672208Z * [new tag] trunk/874069fbe46e82da5cfa405e6c0deb12e89ff608 -> trunk/874069fbe46e82da5cfa405e6c0deb12e89ff608 2025-09-07T07:28:19.0673015Z * [new tag] trunk/8875d6e394da2fffd04f31b28bf258c94d4776a3 -> trunk/8875d6e394da2fffd04f31b28bf258c94d4776a3 2025-09-07T07:28:19.0673786Z * [new tag] trunk/88d94d17e8c5155451393afa6eb3bab48ab61c16 -> trunk/88d94d17e8c5155451393afa6eb3bab48ab61c16 2025-09-07T07:28:19.0674617Z * [new tag] trunk/890626632def7e0ef95a2d01e87a0e4627824a9f -> trunk/890626632def7e0ef95a2d01e87a0e4627824a9f 2025-09-07T07:28:19.0675451Z * [new tag] trunk/8975cda2520b7b1b5bc3b4d8213edf261fa82570 -> trunk/8975cda2520b7b1b5bc3b4d8213edf261fa82570 2025-09-07T07:28:19.0676185Z * [new tag] trunk/89d41d3f61d04f14730ec26f008a59bef6624610 -> trunk/89d41d3f61d04f14730ec26f008a59bef6624610 2025-09-07T07:28:19.0676892Z * [new tag] trunk/8bb213b6d599ef1273fe52f9b1f6d476056c3a41 -> trunk/8bb213b6d599ef1273fe52f9b1f6d476056c3a41 2025-09-07T07:28:19.0677583Z * [new tag] trunk/8e23a1227b5fb2e39afaa7d57c075a75b640a5af -> trunk/8e23a1227b5fb2e39afaa7d57c075a75b640a5af 2025-09-07T07:28:19.0678725Z * [new tag] trunk/8ec551bb354ab2b85fbbba9d461740a20366d248 -> trunk/8ec551bb354ab2b85fbbba9d461740a20366d248 2025-09-07T07:28:19.0679537Z * [new tag] trunk/8fd3c9ce919c8d5c645fd348bba517e948cbc29d -> trunk/8fd3c9ce919c8d5c645fd348bba517e948cbc29d 2025-09-07T07:28:19.0680417Z * [new tag] trunk/90f50f7e68e120d9574e6e3189e37b4280010ad9 -> trunk/90f50f7e68e120d9574e6e3189e37b4280010ad9 2025-09-07T07:28:19.0681181Z * [new tag] trunk/91f0bcf43fc0bc743350d491ac63b77e92054ac9 -> trunk/91f0bcf43fc0bc743350d491ac63b77e92054ac9 2025-09-07T07:28:19.0681979Z * [new tag] trunk/92576a594b8121f6b0b1b5a3ea16d08792fc68ab -> trunk/92576a594b8121f6b0b1b5a3ea16d08792fc68ab 2025-09-07T07:28:19.0682722Z * [new tag] trunk/92a43025e0baa1f2ce345f28d22913b518a1ab9d -> trunk/92a43025e0baa1f2ce345f28d22913b518a1ab9d 2025-09-07T07:28:19.0683305Z * [new tag] trunk/93fb23d6fae7c4e82c4239a1033e522088742634 -> trunk/93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:28:19.0684024Z * [new tag] trunk/9458d1ac3bd70c2af316a8ba95d2c6c9c1199c9c -> trunk/9458d1ac3bd70c2af316a8ba95d2c6c9c1199c9c 2025-09-07T07:28:19.0684935Z * [new tag] trunk/9480cdc0b61488c89a23c2f64f43b2dcedc8728e -> trunk/9480cdc0b61488c89a23c2f64f43b2dcedc8728e 2025-09-07T07:28:19.0685589Z * [new tag] trunk/9491d289b329e4ba4a9f5f5b1be7960671bb7840 -> trunk/9491d289b329e4ba4a9f5f5b1be7960671bb7840 2025-09-07T07:28:19.0686283Z * [new tag] trunk/9499c8761cd2067feb9877414e818f6fd00290f1 -> trunk/9499c8761cd2067feb9877414e818f6fd00290f1 2025-09-07T07:28:19.0687009Z * [new tag] trunk/95ee0bfea99d3d346d6502b91b497d2b35795504 -> trunk/95ee0bfea99d3d346d6502b91b497d2b35795504 2025-09-07T07:28:19.0687717Z * [new tag] trunk/98374612fc2febd686be20761e56bdc2424bc36a -> trunk/98374612fc2febd686be20761e56bdc2424bc36a 2025-09-07T07:28:19.0688612Z * [new tag] trunk/98efc9e93d8fc61eb53cb91378443617cb550500 -> trunk/98efc9e93d8fc61eb53cb91378443617cb550500 2025-09-07T07:28:19.0689462Z * [new tag] trunk/994f2a5dbcbdc915da39bf6f6ce4d1f5e74835c9 -> trunk/994f2a5dbcbdc915da39bf6f6ce4d1f5e74835c9 2025-09-07T07:28:19.0690128Z * [new tag] trunk/99f356fa58c8d726cef022d8710f5491291158f6 -> trunk/99f356fa58c8d726cef022d8710f5491291158f6 2025-09-07T07:28:19.0690888Z * [new tag] trunk/9a1c5c0a078b94d13ac5c1ae0d754d19fb73bf99 -> trunk/9a1c5c0a078b94d13ac5c1ae0d754d19fb73bf99 2025-09-07T07:28:19.0691610Z * [new tag] trunk/9a665ca3c472384e9d722bddba79e5a7680f1abd -> trunk/9a665ca3c472384e9d722bddba79e5a7680f1abd 2025-09-07T07:28:19.0692396Z * [new tag] trunk/9aedb3cd87b52160872173c177f61053d97bed57 -> trunk/9aedb3cd87b52160872173c177f61053d97bed57 2025-09-07T07:28:19.0693147Z * [new tag] trunk/9b81fe281da41f2421506339d26b027a468902f4 -> trunk/9b81fe281da41f2421506339d26b027a468902f4 2025-09-07T07:28:19.0693847Z * [new tag] trunk/9bdcee01f86e2969cff1140cdecfca13cb51816e -> trunk/9bdcee01f86e2969cff1140cdecfca13cb51816e 2025-09-07T07:28:19.0694603Z * [new tag] trunk/9c03d6be87eedc06e524e202e07a7e776551a839 -> trunk/9c03d6be87eedc06e524e202e07a7e776551a839 2025-09-07T07:28:19.0695294Z * [new tag] trunk/9c957723a0fedd9c637e63e023a613019e2cab60 -> trunk/9c957723a0fedd9c637e63e023a613019e2cab60 2025-09-07T07:28:19.0696026Z * [new tag] trunk/9e5247f51d81735e5f1e65e80588985fa93bccc5 -> trunk/9e5247f51d81735e5f1e65e80588985fa93bccc5 2025-09-07T07:28:19.0696839Z * [new tag] trunk/9eadb37cdd699f7e8e8177a5227bfeb16184ef26 -> trunk/9eadb37cdd699f7e8e8177a5227bfeb16184ef26 2025-09-07T07:28:19.0697561Z * [new tag] trunk/a00cdc1e4159db73c9ffb3f25e93e55877709a29 -> trunk/a00cdc1e4159db73c9ffb3f25e93e55877709a29 2025-09-07T07:28:19.0698290Z * [new tag] trunk/a02ee4a816d11380c6f564c1aba64d56af5ba705 -> trunk/a02ee4a816d11380c6f564c1aba64d56af5ba705 2025-09-07T07:28:19.0699000Z * [new tag] trunk/a3c7f77e50f900721817934120d60c2361b3c40d -> trunk/a3c7f77e50f900721817934120d60c2361b3c40d 2025-09-07T07:28:19.0699746Z * [new tag] trunk/a3d72b09ae12126a2b7d4a63a45ac100a882a802 -> trunk/a3d72b09ae12126a2b7d4a63a45ac100a882a802 2025-09-07T07:28:19.0700486Z * [new tag] trunk/a3e5466002791da609fcb069155d8ee347baee92 -> trunk/a3e5466002791da609fcb069155d8ee347baee92 2025-09-07T07:28:19.0701256Z * [new tag] trunk/a714437093ed196eee28f7de454cf4c41badc098 -> trunk/a714437093ed196eee28f7de454cf4c41badc098 2025-09-07T07:28:19.0701978Z * [new tag] trunk/a75e8cd27098f290de0b7439685d05ce02e91356 -> trunk/a75e8cd27098f290de0b7439685d05ce02e91356 2025-09-07T07:28:19.0702603Z * [new tag] trunk/a8d6943d36c1c2a5f90d3573460695bad4b623ae -> trunk/a8d6943d36c1c2a5f90d3573460695bad4b623ae 2025-09-07T07:28:19.0703377Z * [new tag] trunk/a918bbad6ab20649ff82eefb48417ecbe96bcb34 -> trunk/a918bbad6ab20649ff82eefb48417ecbe96bcb34 2025-09-07T07:28:19.0704146Z * [new tag] trunk/a99d8d39bc842d6ebc3e368b178e4884d24b056e -> trunk/a99d8d39bc842d6ebc3e368b178e4884d24b056e 2025-09-07T07:28:19.0704807Z * [new tag] trunk/aac1a50a191b4102d566c9c1ea22f06d6c2e3f02 -> trunk/aac1a50a191b4102d566c9c1ea22f06d6c2e3f02 2025-09-07T07:28:19.0705553Z * [new tag] trunk/aad96a202244c7d0d120c04ba8db593edd8c0f92 -> trunk/aad96a202244c7d0d120c04ba8db593edd8c0f92 2025-09-07T07:28:19.0706310Z * [new tag] trunk/ab643e4dbbaf7b663d4237514cbf01af9b11565c -> trunk/ab643e4dbbaf7b663d4237514cbf01af9b11565c 2025-09-07T07:28:19.0707011Z * [new tag] trunk/abc447174cd2cf8591edbc70a9f836f9a5779f47 -> trunk/abc447174cd2cf8591edbc70a9f836f9a5779f47 2025-09-07T07:28:19.0707784Z * [new tag] trunk/acece97c3a9dceb63194e314da93fdf37cf15a0d -> trunk/acece97c3a9dceb63194e314da93fdf37cf15a0d 2025-09-07T07:28:19.0708588Z * [new tag] trunk/adae7f66aacf3f248c3101b858cf98d5809119fa -> trunk/adae7f66aacf3f248c3101b858cf98d5809119fa 2025-09-07T07:28:19.0709363Z * [new tag] trunk/ae0edc133e61e3b16caf0b2ee0ff3f33ab72af4c -> trunk/ae0edc133e61e3b16caf0b2ee0ff3f33ab72af4c 2025-09-07T07:28:19.0710117Z * [new tag] trunk/aed33a8fcbd60b052d4559d261390c5797129c6d -> trunk/aed33a8fcbd60b052d4559d261390c5797129c6d 2025-09-07T07:28:19.0710963Z * [new tag] trunk/b04e922712080a3652e438d05e8bb74e0cd2d238 -> trunk/b04e922712080a3652e438d05e8bb74e0cd2d238 2025-09-07T07:28:19.0711785Z * [new tag] trunk/b0a3e58dd71c1a039ac0ef51e5bd8f704f632f6f -> trunk/b0a3e58dd71c1a039ac0ef51e5bd8f704f632f6f 2025-09-07T07:28:19.0712539Z * [new tag] trunk/b16d3f4c8c01d461c2f01064e9ca5fa2b33f5cf1 -> trunk/b16d3f4c8c01d461c2f01064e9ca5fa2b33f5cf1 2025-09-07T07:28:19.0713331Z * [new tag] trunk/b18bb6796f210a183e687d9d64984a5a9d13cf09 -> trunk/b18bb6796f210a183e687d9d64984a5a9d13cf09 2025-09-07T07:28:19.0714126Z * [new tag] trunk/b1bb98ddebdd3e41bf7987372409bdce96ae55de -> trunk/b1bb98ddebdd3e41bf7987372409bdce96ae55de 2025-09-07T07:28:19.0714781Z * [new tag] trunk/b2b4add0e754411372060e1d7b4057a66439172b -> trunk/b2b4add0e754411372060e1d7b4057a66439172b 2025-09-07T07:28:19.0715552Z * [new tag] trunk/b2c7b9ad2dc5a7c0b61febd307761bd5bc2f0f05 -> trunk/b2c7b9ad2dc5a7c0b61febd307761bd5bc2f0f05 2025-09-07T07:28:19.0716266Z * [new tag] trunk/b40d9432be44a6b5974ee62e7d19c3c61c5ece37 -> trunk/b40d9432be44a6b5974ee62e7d19c3c61c5ece37 2025-09-07T07:28:19.0717029Z * [new tag] trunk/b4ad38279b178b7bd14355123c1101e2e853e77b -> trunk/b4ad38279b178b7bd14355123c1101e2e853e77b 2025-09-07T07:28:19.0717763Z * [new tag] trunk/b67c41039835bd9b20b83cd6233e86baaa5f5dde -> trunk/b67c41039835bd9b20b83cd6233e86baaa5f5dde 2025-09-07T07:28:19.0718678Z * [new tag] trunk/b6d0a9ea9056ede4f7024dbf3bd6c43be3aff49c -> trunk/b6d0a9ea9056ede4f7024dbf3bd6c43be3aff49c 2025-09-07T07:28:19.0719435Z * [new tag] trunk/b7dad7dd49448c88d0751fa2e29c70afe985f734 -> trunk/b7dad7dd49448c88d0751fa2e29c70afe985f734 2025-09-07T07:28:19.0720406Z * [new tag] trunk/b7e207ca9f046ddd716076965a0cce403ba99052 -> trunk/b7e207ca9f046ddd716076965a0cce403ba99052 2025-09-07T07:28:19.0721220Z * [new tag] trunk/b919560c4a7010e2d89facee25586269a994746e -> trunk/b919560c4a7010e2d89facee25586269a994746e 2025-09-07T07:28:19.0722031Z * [new tag] trunk/b9ba612f7a968f7b27e121ca8f4d0a4d954f5354 -> trunk/b9ba612f7a968f7b27e121ca8f4d0a4d954f5354 2025-09-07T07:28:19.0722862Z * [new tag] trunk/ba7f546ccccb5e0b36d9070dc25f26a9647f89f8 -> trunk/ba7f546ccccb5e0b36d9070dc25f26a9647f89f8 2025-09-07T07:28:19.0723560Z * [new tag] trunk/bb950284c7e72905994bc25dd436c10e48088d85 -> trunk/bb950284c7e72905994bc25dd436c10e48088d85 2025-09-07T07:28:19.0724392Z * [new tag] trunk/bbedc71fd3267c639c38b4ec25eaa22f973d9c4d -> trunk/bbedc71fd3267c639c38b4ec25eaa22f973d9c4d 2025-09-07T07:28:19.0725074Z * [new tag] trunk/bc4db2c27fce6ff1648bdc5af31ec225d2a31f37 -> trunk/bc4db2c27fce6ff1648bdc5af31ec225d2a31f37 2025-09-07T07:28:19.0725718Z * [new tag] trunk/bc505977fb66677a09c31155c987330fbb18a865 -> trunk/bc505977fb66677a09c31155c987330fbb18a865 2025-09-07T07:28:19.0726508Z * [new tag] trunk/bd39e47feea7326afb5bbb67fcb1e69279239527 -> trunk/bd39e47feea7326afb5bbb67fcb1e69279239527 2025-09-07T07:28:19.0727343Z * [new tag] trunk/be5b03dde96638f25ffd732a4fed7e41b4cf40e1 -> trunk/be5b03dde96638f25ffd732a4fed7e41b4cf40e1 2025-09-07T07:28:19.0728083Z * [new tag] trunk/bffc7dd1f374d8408911cd22c6b3d6df39ded9b3 -> trunk/bffc7dd1f374d8408911cd22c6b3d6df39ded9b3 2025-09-07T07:28:19.0728914Z * [new tag] trunk/c024b1f5a18d5c5aee5cc2acdd4c52b24b93ffcf -> trunk/c024b1f5a18d5c5aee5cc2acdd4c52b24b93ffcf 2025-09-07T07:28:19.0729621Z * [new tag] trunk/c0983e6cc0acf71689e1851d12609e00b3f59371 -> trunk/c0983e6cc0acf71689e1851d12609e00b3f59371 2025-09-07T07:28:19.0730334Z * [new tag] trunk/c10195e723eeeedd099ed8b73eda7184ca618fad -> trunk/c10195e723eeeedd099ed8b73eda7184ca618fad 2025-09-07T07:28:19.0731100Z * [new tag] trunk/c157cf6488ade6a7ee2ce2d25b059e1335630a99 -> trunk/c157cf6488ade6a7ee2ce2d25b059e1335630a99 2025-09-07T07:28:19.0731833Z * [new tag] trunk/c2a30246172fd71d56529907ffd3c27b76b1f3a7 -> trunk/c2a30246172fd71d56529907ffd3c27b76b1f3a7 2025-09-07T07:28:19.0732593Z * [new tag] trunk/c32111149921b48bfef909293f1049e21619ed76 -> trunk/c32111149921b48bfef909293f1049e21619ed76 2025-09-07T07:28:19.0733258Z * [new tag] trunk/c37103234afc832dcad307e9016230810957c9d5 -> trunk/c37103234afc832dcad307e9016230810957c9d5 2025-09-07T07:28:19.0733974Z * [new tag] trunk/c3ceca2995cd35e1376c4b0704669bff1a81e836 -> trunk/c3ceca2995cd35e1376c4b0704669bff1a81e836 2025-09-07T07:28:19.0734740Z * [new tag] trunk/c3d54dea9febb1236d48d19e5d4876a63f2e20fd -> trunk/c3d54dea9febb1236d48d19e5d4876a63f2e20fd 2025-09-07T07:28:19.0735456Z * [new tag] trunk/c465b3d52c5687fe910d35a5c75341b77f821741 -> trunk/c465b3d52c5687fe910d35a5c75341b77f821741 2025-09-07T07:28:19.0736210Z * [new tag] trunk/c5b8a10be5e89396da916d1069ffcb7135f0372b -> trunk/c5b8a10be5e89396da916d1069ffcb7135f0372b 2025-09-07T07:28:19.0736871Z * [new tag] trunk/c7e41071a08f4045bc11ab60ec366d7357d56e30 -> trunk/c7e41071a08f4045bc11ab60ec366d7357d56e30 2025-09-07T07:28:19.0737784Z * [new tag] trunk/c98ddaca6d2e19ca37aff00c4ff0cda1e9a6ff65 -> trunk/c98ddaca6d2e19ca37aff00c4ff0cda1e9a6ff65 2025-09-07T07:28:19.0738537Z * [new tag] trunk/cb1e31362c7b53acf4ac95b9f8878064c184f03b -> trunk/cb1e31362c7b53acf4ac95b9f8878064c184f03b 2025-09-07T07:28:19.0739312Z * [new tag] trunk/cbfb005f7cce79974795b148e265f594f59477c8 -> trunk/cbfb005f7cce79974795b148e265f594f59477c8 2025-09-07T07:28:19.0740076Z * [new tag] trunk/cc5bdd12401bda835291d2f3cb297132ebdbf358 -> trunk/cc5bdd12401bda835291d2f3cb297132ebdbf358 2025-09-07T07:28:19.0740989Z * [new tag] trunk/cd529b686d54bbaa443f5b310140de48422d96c7 -> trunk/cd529b686d54bbaa443f5b310140de48422d96c7 2025-09-07T07:28:19.0741687Z * [new tag] trunk/cec0ff122815582af5302360aff03676558c5c87 -> trunk/cec0ff122815582af5302360aff03676558c5c87 2025-09-07T07:28:19.0742419Z * [new tag] trunk/d11720efdb563d02cf4f7d324311fb15a755268e -> trunk/d11720efdb563d02cf4f7d324311fb15a755268e 2025-09-07T07:28:19.0743140Z * [new tag] trunk/d1706d9128ae24d9048167e80d3fe5196d19035e -> trunk/d1706d9128ae24d9048167e80d3fe5196d19035e 2025-09-07T07:28:19.0743972Z * [new tag] trunk/d1a15abfdcaef138f2d9e93a9f46be44f30b766d -> trunk/d1a15abfdcaef138f2d9e93a9f46be44f30b766d 2025-09-07T07:28:19.0744877Z * [new tag] trunk/d232a95d4a79404ca05c1f52d37fde7339dcdf49 -> trunk/d232a95d4a79404ca05c1f52d37fde7339dcdf49 2025-09-07T07:28:19.0745589Z * [new tag] trunk/d2d4c8e9b2371c9aacfb771d9402ac7427b9778e -> trunk/d2d4c8e9b2371c9aacfb771d9402ac7427b9778e 2025-09-07T07:28:19.0746316Z * [new tag] trunk/d33840c542b387ab08ba49aa6c45aa9567fd9be7 -> trunk/d33840c542b387ab08ba49aa6c45aa9567fd9be7 2025-09-07T07:28:19.0747050Z * [new tag] trunk/d5643e8f3a648a99636bfa1f2a41d54bd3c0d0f1 -> trunk/d5643e8f3a648a99636bfa1f2a41d54bd3c0d0f1 2025-09-07T07:28:19.0747818Z * [new tag] trunk/d5b38410b5b6cf75c7a7389972777a6497926ee7 -> trunk/d5b38410b5b6cf75c7a7389972777a6497926ee7 2025-09-07T07:28:19.0748396Z * [new tag] trunk/d5e0f4202ba14632e4d14862ace096609e763462 -> trunk/d5e0f4202ba14632e4d14862ace096609e763462 2025-09-07T07:28:19.0749279Z * [new tag] trunk/d636c181f9140a7b59be10b36eae23039fc2bb72 -> trunk/d636c181f9140a7b59be10b36eae23039fc2bb72 2025-09-07T07:28:19.0750484Z * [new tag] trunk/d64718503728001a1e78168fd7f2d4ff23e57285 -> trunk/d64718503728001a1e78168fd7f2d4ff23e57285 2025-09-07T07:28:19.0751316Z * [new tag] trunk/d67c29ad22670320d676b02e394274af34e8e643 -> trunk/d67c29ad22670320d676b02e394274af34e8e643 2025-09-07T07:28:19.0752058Z * [new tag] trunk/d6b74568e2c98ce58ecc145b72ac66d4caf7ce95 -> trunk/d6b74568e2c98ce58ecc145b72ac66d4caf7ce95 2025-09-07T07:28:19.0752824Z * [new tag] trunk/d711f27845abd45007ccab6076649ebd896c2661 -> trunk/d711f27845abd45007ccab6076649ebd896c2661 2025-09-07T07:28:19.0753561Z * [new tag] trunk/d9d6dde0f42d4bcc8c97671ac50d5096c7e500ab -> trunk/d9d6dde0f42d4bcc8c97671ac50d5096c7e500ab 2025-09-07T07:28:19.0754334Z * [new tag] trunk/da4db4b33d1fdd046650cf19fdbac581a19bf2f9 -> trunk/da4db4b33d1fdd046650cf19fdbac581a19bf2f9 2025-09-07T07:28:19.0754975Z * [new tag] trunk/dac8a4b91c01c3bbc96f54e621b1ea4ffdbd29d1 -> trunk/dac8a4b91c01c3bbc96f54e621b1ea4ffdbd29d1 2025-09-07T07:28:19.0755744Z * [new tag] trunk/dbec08729fb9848bebed6048c63831b87170d061 -> trunk/dbec08729fb9848bebed6048c63831b87170d061 2025-09-07T07:28:19.0756400Z * [new tag] trunk/dcf385395d838f38c8dca25913578230dd43099a -> trunk/dcf385395d838f38c8dca25913578230dd43099a 2025-09-07T07:28:19.0757129Z * [new tag] trunk/dd2519abe83ec3c40d4797492434e41fe3b47e17 -> trunk/dd2519abe83ec3c40d4797492434e41fe3b47e17 2025-09-07T07:28:19.0757897Z * [new tag] trunk/dec72ea4b006dd0fbcaaaa106ad273d73807ab9d -> trunk/dec72ea4b006dd0fbcaaaa106ad273d73807ab9d 2025-09-07T07:28:19.0758606Z * [new tag] trunk/e0a62b266c021b910ce6dc02a6c9429210487717 -> trunk/e0a62b266c021b910ce6dc02a6c9429210487717 2025-09-07T07:28:19.0759390Z * [new tag] trunk/e19e02c84c9dcc408375e5cae3b0709c18b99228 -> trunk/e19e02c84c9dcc408375e5cae3b0709c18b99228 2025-09-07T07:28:19.0760167Z * [new tag] trunk/e304ea4e69d3a7deeb7e48c7450c214a4c953937 -> trunk/e304ea4e69d3a7deeb7e48c7450c214a4c953937 2025-09-07T07:28:19.0760941Z * [new tag] trunk/e3068cdb446adefb5a875616ba37a60235391439 -> trunk/e3068cdb446adefb5a875616ba37a60235391439 2025-09-07T07:28:19.0761732Z * [new tag] trunk/e381d4b0205d5f126c1de534f867ba776f7c3ee6 -> trunk/e381d4b0205d5f126c1de534f867ba776f7c3ee6 2025-09-07T07:28:19.0762556Z * [new tag] trunk/e4bd0ff4f8981b805df32ea5b3550621965ea4f2 -> trunk/e4bd0ff4f8981b805df32ea5b3550621965ea4f2 2025-09-07T07:28:19.0763195Z * [new tag] trunk/e532c9d4f1cdcbc1ea9628f55b9813e77847bdc7 -> trunk/e532c9d4f1cdcbc1ea9628f55b9813e77847bdc7 2025-09-07T07:28:19.0763933Z * [new tag] trunk/e92cd9415377403b6e90585e764639e2e0b5973b -> trunk/e92cd9415377403b6e90585e764639e2e0b5973b 2025-09-07T07:28:19.0764715Z * [new tag] trunk/e9481b6617b5576b099d8ca5798111592e9ad090 -> trunk/e9481b6617b5576b099d8ca5798111592e9ad090 2025-09-07T07:28:19.0765330Z * [new tag] trunk/ea1883dfd3e42defe37b11202b878bb76defa087 -> trunk/ea1883dfd3e42defe37b11202b878bb76defa087 2025-09-07T07:28:19.0766150Z * [new tag] trunk/eac3d6f04cfbbebe3d470dacd216da7d4b1f95a8 -> trunk/eac3d6f04cfbbebe3d470dacd216da7d4b1f95a8 2025-09-07T07:28:19.0766845Z * [new tag] trunk/eb18d32bda75189494d955aa001ade15f10333de -> trunk/eb18d32bda75189494d955aa001ade15f10333de 2025-09-07T07:28:19.0767480Z * [new tag] trunk/ef3be6726f7ff4b77c22db10cec5b686f9107ea9 -> trunk/ef3be6726f7ff4b77c22db10cec5b686f9107ea9 2025-09-07T07:28:19.0768232Z * [new tag] trunk/ef8aabd42422725026cb4dbf48aafa9efa226a04 -> trunk/ef8aabd42422725026cb4dbf48aafa9efa226a04 2025-09-07T07:28:19.0769673Z * [new tag] trunk/f00445b43eee57e20bb9316fa796ca23bf73373b -> trunk/f00445b43eee57e20bb9316fa796ca23bf73373b 2025-09-07T07:28:19.0770464Z * [new tag] trunk/f0c391102b754e3b145e8c59231d2df563487e37 -> trunk/f0c391102b754e3b145e8c59231d2df563487e37 2025-09-07T07:28:19.0771286Z * [new tag] trunk/f27985b7e796fb66a1b476284ba42d8cb360a751 -> trunk/f27985b7e796fb66a1b476284ba42d8cb360a751 2025-09-07T07:28:19.0772096Z * [new tag] trunk/f36f285953700f971552083a5da9d0ceacb63bbd -> trunk/f36f285953700f971552083a5da9d0ceacb63bbd 2025-09-07T07:28:19.0772877Z * [new tag] trunk/f3cebec39ebc110e1c8b06e741896585f7892dbb -> trunk/f3cebec39ebc110e1c8b06e741896585f7892dbb 2025-09-07T07:28:19.0773524Z * [new tag] trunk/f4c33cd44acac92c0b451a04da20ebe9370e5b0c -> trunk/f4c33cd44acac92c0b451a04da20ebe9370e5b0c 2025-09-07T07:28:19.0774281Z * [new tag] trunk/f612045ce105f008b2b675e2fc870163babeb2e8 -> trunk/f612045ce105f008b2b675e2fc870163babeb2e8 2025-09-07T07:28:19.0775093Z * [new tag] trunk/f8746b878dfc1e9639d42cbde832e9b9e792c86c -> trunk/f8746b878dfc1e9639d42cbde832e9b9e792c86c 2025-09-07T07:28:19.0775762Z * [new tag] trunk/f8ffa9194e26523e5f976d4a824d5cc58922727c -> trunk/f8ffa9194e26523e5f976d4a824d5cc58922727c 2025-09-07T07:28:19.0776505Z * [new tag] trunk/f981a7fa5230b98974291fdde32fe8488bc5d469 -> trunk/f981a7fa5230b98974291fdde32fe8488bc5d469 2025-09-07T07:28:19.0777247Z * [new tag] trunk/fbf3d2027daabbcb44d0af274b139be2a248a4f7 -> trunk/fbf3d2027daabbcb44d0af274b139be2a248a4f7 2025-09-07T07:28:19.0778162Z * [new tag] trunk/fca2601c9d628e1bd2d75c7318cd22c4e8c832aa -> trunk/fca2601c9d628e1bd2d75c7318cd22c4e8c832aa 2025-09-07T07:28:19.0778949Z * [new tag] trunk/fea20775ad96bdca972a1811d7d3372f368614ab -> trunk/fea20775ad96bdca972a1811d7d3372f368614ab 2025-09-07T07:28:19.0779561Z * [new tag] trunk/fefee081642f87419a21dc852f7167d4640443cd -> trunk/fefee081642f87419a21dc852f7167d4640443cd 2025-09-07T07:28:19.0780131Z * [new tag] v0.1.1 -> v0.1.1 2025-09-07T07:28:19.0780805Z * [new tag] v0.1.10 -> v0.1.10 2025-09-07T07:28:19.0781452Z * [new tag] v0.1.11 -> v0.1.11 2025-09-07T07:28:19.0782140Z * [new tag] v0.1.12 -> v0.1.12 2025-09-07T07:28:19.0782772Z * [new tag] v0.1.2 -> v0.1.2 2025-09-07T07:28:19.0783439Z * [new tag] v0.1.3 -> v0.1.3 2025-09-07T07:28:19.0784109Z * [new tag] v0.1.4 -> v0.1.4 2025-09-07T07:28:19.0784733Z * [new tag] v0.1.5 -> v0.1.5 2025-09-07T07:28:19.0785444Z * [new tag] v0.1.6 -> v0.1.6 2025-09-07T07:28:19.0786081Z * [new tag] v0.1.7 -> v0.1.7 2025-09-07T07:28:19.0786828Z * [new tag] v0.1.8 -> v0.1.8 2025-09-07T07:28:19.0787600Z * [new tag] v0.1.9 -> v0.1.9 2025-09-07T07:28:19.0788154Z * [new tag] v0.2.0 -> v0.2.0 2025-09-07T07:28:19.0788821Z * [new tag] v0.3.0 -> v0.3.0 2025-09-07T07:28:19.0789624Z * [new tag] v0.3.1 -> v0.3.1 2025-09-07T07:28:19.0790295Z * [new tag] v0.4.0 -> v0.4.0 2025-09-07T07:28:19.0790928Z * [new tag] v0.4.1 -> v0.4.1 2025-09-07T07:28:19.0791596Z * [new tag] v1.0.0 -> v1.0.0 2025-09-07T07:28:19.0792331Z * [new tag] v1.0.0a0 -> v1.0.0a0 2025-09-07T07:28:19.0792978Z * [new tag] v1.0.1 -> v1.0.1 2025-09-07T07:28:19.0793676Z * [new tag] v1.0rc0 -> v1.0rc0 2025-09-07T07:28:19.0794204Z * [new tag] v1.0rc1 -> v1.0rc1 2025-09-07T07:28:19.0794893Z * [new tag] v1.1.0 -> v1.1.0 2025-09-07T07:28:19.0795637Z * [new tag] v1.1.0a0 -> v1.1.0a0 2025-09-07T07:28:19.0796552Z * [new tag] v1.10.0 -> v1.10.0 2025-09-07T07:28:19.0797296Z * [new tag] v1.10.0-rc1 -> v1.10.0-rc1 2025-09-07T07:28:19.0798070Z * [new tag] v1.10.0-rc2 -> v1.10.0-rc2 2025-09-07T07:28:19.0798558Z * [new tag] v1.10.0-rc3 -> v1.10.0-rc3 2025-09-07T07:28:19.0799278Z * [new tag] v1.10.1 -> v1.10.1 2025-09-07T07:28:19.0799857Z * [new tag] v1.10.1-rc1 -> v1.10.1-rc1 2025-09-07T07:28:19.0800386Z * [new tag] v1.10.2 -> v1.10.2 2025-09-07T07:28:19.0800908Z * [new tag] v1.10.2-rc1 -> v1.10.2-rc1 2025-09-07T07:28:19.0801594Z * [new tag] v1.11.0 -> v1.11.0 2025-09-07T07:28:19.0802337Z * [new tag] v1.11.0-rc1 -> v1.11.0-rc1 2025-09-07T07:28:19.0803127Z * [new tag] v1.11.0-rc2 -> v1.11.0-rc2 2025-09-07T07:28:19.0803946Z * [new tag] v1.11.0-rc3 -> v1.11.0-rc3 2025-09-07T07:28:19.0804654Z * [new tag] v1.11.0-rc4 -> v1.11.0-rc4 2025-09-07T07:28:19.0805434Z * [new tag] v1.11.0-rc5 -> v1.11.0-rc5 2025-09-07T07:28:19.0806012Z * [new tag] v1.11.0-rc6 -> v1.11.0-rc6 2025-09-07T07:28:19.0806540Z * [new tag] v1.11.0-rc7 -> v1.11.0-rc7 2025-09-07T07:28:19.0807217Z * [new tag] v1.12.0 -> v1.12.0 2025-09-07T07:28:19.0808002Z * [new tag] v1.12.0-rc1 -> v1.12.0-rc1 2025-09-07T07:28:19.0808947Z * [new tag] v1.12.0-rc2 -> v1.12.0-rc2 2025-09-07T07:28:19.0809814Z * [new tag] v1.12.0-rc3 -> v1.12.0-rc3 2025-09-07T07:28:19.0810505Z * [new tag] v1.12.0-rc4 -> v1.12.0-rc4 2025-09-07T07:28:19.0811196Z * [new tag] v1.12.0-rc5 -> v1.12.0-rc5 2025-09-07T07:28:19.0812061Z * [new tag] v1.12.0-rc6 -> v1.12.0-rc6 2025-09-07T07:28:19.0812569Z * [new tag] v1.12.0-rc7 -> v1.12.0-rc7 2025-09-07T07:28:19.0813129Z * [new tag] v1.12.0-rc8 -> v1.12.0-rc8 2025-09-07T07:28:19.0813666Z * [new tag] v1.12.1 -> v1.12.1 2025-09-07T07:28:19.0814478Z * [new tag] v1.12.1-rc1 -> v1.12.1-rc1 2025-09-07T07:28:19.0815171Z * [new tag] v1.12.1-rc2 -> v1.12.1-rc2 2025-09-07T07:28:19.0815978Z * [new tag] v1.12.1-rc3 -> v1.12.1-rc3 2025-09-07T07:28:19.0816850Z * [new tag] v1.12.1-rc4 -> v1.12.1-rc4 2025-09-07T07:28:19.0817260Z * [new tag] v1.12.1-rc5 -> v1.12.1-rc5 2025-09-07T07:28:19.0817957Z * [new tag] v1.13.0 -> v1.13.0 2025-09-07T07:28:19.0818638Z * [new tag] v1.13.0-rc1 -> v1.13.0-rc1 2025-09-07T07:28:19.0819358Z * [new tag] v1.13.0-rc2 -> v1.13.0-rc2 2025-09-07T07:28:19.0820061Z * [new tag] v1.13.0-rc3 -> v1.13.0-rc3 2025-09-07T07:28:19.0820901Z * [new tag] v1.13.0-rc4 -> v1.13.0-rc4 2025-09-07T07:28:19.0821501Z * [new tag] v1.13.0-rc5 -> v1.13.0-rc5 2025-09-07T07:28:19.0821982Z * [new tag] v1.13.0-rc6 -> v1.13.0-rc6 2025-09-07T07:28:19.0822744Z * [new tag] v1.13.1 -> v1.13.1 2025-09-07T07:28:19.0823255Z * [new tag] v1.13.1-rc1 -> v1.13.1-rc1 2025-09-07T07:28:19.0823982Z * [new tag] v1.2.0 -> v1.2.0 2025-09-07T07:28:19.0824708Z * [new tag] v1.2.0a0 -> v1.2.0a0 2025-09-07T07:28:19.0825361Z * [new tag] v1.3.0 -> v1.3.0 2025-09-07T07:28:19.0826120Z * [new tag] v1.3.0a0 -> v1.3.0a0 2025-09-07T07:28:19.0826642Z * [new tag] v1.3.1 -> v1.3.1 2025-09-07T07:28:19.0827275Z * [new tag] v1.4.0 -> v1.4.0 2025-09-07T07:28:19.0828015Z * [new tag] v1.4.0a0 -> v1.4.0a0 2025-09-07T07:28:19.0828508Z * [new tag] v1.4.1 -> v1.4.1 2025-09-07T07:28:19.0829324Z * [new tag] v1.5.0 -> v1.5.0 2025-09-07T07:28:19.0830169Z * [new tag] v1.5.0-rc1 -> v1.5.0-rc1 2025-09-07T07:28:19.0830891Z * [new tag] v1.5.0-rc2 -> v1.5.0-rc2 2025-09-07T07:28:19.0831684Z * [new tag] v1.5.0-rc3 -> v1.5.0-rc3 2025-09-07T07:28:19.0832274Z * [new tag] v1.5.0-rc4 -> v1.5.0-rc4 2025-09-07T07:28:19.0832843Z * [new tag] v1.5.0-rc5 -> v1.5.0-rc5 2025-09-07T07:28:19.0833579Z * [new tag] v1.5.1 -> v1.5.1 2025-09-07T07:28:19.0834164Z * [new tag] v1.5.1-rc1 -> v1.5.1-rc1 2025-09-07T07:28:19.0834658Z * [new tag] v1.6.0 -> v1.6.0 2025-09-07T07:28:19.0835413Z * [new tag] v1.6.0-rc1 -> v1.6.0-rc1 2025-09-07T07:28:19.0836212Z * [new tag] v1.6.0-rc2 -> v1.6.0-rc2 2025-09-07T07:28:19.0836955Z * [new tag] v1.6.0-rc3 -> v1.6.0-rc3 2025-09-07T07:28:19.0837684Z * [new tag] v1.6.0-rc4 -> v1.6.0-rc4 2025-09-07T07:28:19.0838379Z * [new tag] v1.6.0-rc5 -> v1.6.0-rc5 2025-09-07T07:28:19.0839062Z * [new tag] v1.6.0-rc6 -> v1.6.0-rc6 2025-09-07T07:28:19.0839568Z * [new tag] v1.6.0-rc7 -> v1.6.0-rc7 2025-09-07T07:28:19.0840303Z * [new tag] v1.7.0 -> v1.7.0 2025-09-07T07:28:19.0841115Z * [new tag] v1.7.0-rc1 -> v1.7.0-rc1 2025-09-07T07:28:19.0841878Z * [new tag] v1.7.0-rc2 -> v1.7.0-rc2 2025-09-07T07:28:19.0842620Z * [new tag] v1.7.0-rc3 -> v1.7.0-rc3 2025-09-07T07:28:19.0843121Z * [new tag] v1.7.0-rc4 -> v1.7.0-rc4 2025-09-07T07:28:19.0843855Z * [new tag] v1.7.1 -> v1.7.1 2025-09-07T07:28:19.0844791Z * [new tag] v1.7.1-rc1 -> v1.7.1-rc1 2025-09-07T07:28:19.0845436Z * [new tag] v1.7.1-rc2 -> v1.7.1-rc2 2025-09-07T07:28:19.0845975Z * [new tag] v1.7.1-rc3 -> v1.7.1-rc3 2025-09-07T07:28:19.0846710Z * [new tag] v1.8.0 -> v1.8.0 2025-09-07T07:28:19.0847257Z * [new tag] v1.8.0-rc1 -> v1.8.0-rc1 2025-09-07T07:28:19.0847962Z * [new tag] v1.8.0-rc2 -> v1.8.0-rc2 2025-09-07T07:28:19.0848699Z * [new tag] v1.8.0-rc3 -> v1.8.0-rc3 2025-09-07T07:28:19.0849429Z * [new tag] v1.8.0-rc4 -> v1.8.0-rc4 2025-09-07T07:28:19.0849984Z * [new tag] v1.8.0-rc5 -> v1.8.0-rc5 2025-09-07T07:28:19.0850502Z * [new tag] v1.8.1 -> v1.8.1 2025-09-07T07:28:19.0851195Z * [new tag] v1.8.1-rc1 -> v1.8.1-rc1 2025-09-07T07:28:19.0851782Z * [new tag] v1.8.1-rc2 -> v1.8.1-rc2 2025-09-07T07:28:19.0852495Z * [new tag] v1.8.1-rc3 -> v1.8.1-rc3 2025-09-07T07:28:19.0853701Z * [new tag] v1.8.2 -> v1.8.2 2025-09-07T07:28:19.0854232Z * [new tag] v1.8.2-rc1 -> v1.8.2-rc1 2025-09-07T07:28:19.0854988Z * [new tag] v1.9.0 -> v1.9.0 2025-09-07T07:28:19.0855682Z * [new tag] v1.9.0-rc1 -> v1.9.0-rc1 2025-09-07T07:28:19.0856538Z * [new tag] v1.9.0-rc2 -> v1.9.0-rc2 2025-09-07T07:28:19.0857320Z * [new tag] v1.9.0-rc3 -> v1.9.0-rc3 2025-09-07T07:28:19.0857873Z * [new tag] v1.9.0-rc4 -> v1.9.0-rc4 2025-09-07T07:28:19.0858629Z * [new tag] v1.9.1 -> v1.9.1 2025-09-07T07:28:19.0859508Z * [new tag] v1.9.1-rc1 -> v1.9.1-rc1 2025-09-07T07:28:19.0860013Z * [new tag] v1.9.1-rc2 -> v1.9.1-rc2 2025-09-07T07:28:19.0860768Z * [new tag] v2.0.0 -> v2.0.0 2025-09-07T07:28:19.0861452Z * [new tag] v2.0.0-rc1 -> v2.0.0-rc1 2025-09-07T07:28:19.0862244Z * [new tag] v2.0.0-rc2 -> v2.0.0-rc2 2025-09-07T07:28:19.0863069Z * [new tag] v2.0.0-rc3 -> v2.0.0-rc3 2025-09-07T07:28:19.0863791Z * [new tag] v2.0.0-rc4 -> v2.0.0-rc4 2025-09-07T07:28:19.0864478Z * [new tag] v2.0.0-rc5 -> v2.0.0-rc5 2025-09-07T07:28:19.0865066Z * [new tag] v2.0.0-rc6 -> v2.0.0-rc6 2025-09-07T07:28:19.0868474Z * [new tag] v2.0.1 -> v2.0.1 2025-09-07T07:28:19.0869331Z * [new tag] v2.0.1-rc1 -> v2.0.1-rc1 2025-09-07T07:28:19.0869856Z * [new tag] v2.0.1-rc2 -> v2.0.1-rc2 2025-09-07T07:28:19.0870528Z * [new tag] v2.0.1-rc3 -> v2.0.1-rc3 2025-09-07T07:28:19.0871019Z * [new tag] v2.0.1-rc4 -> v2.0.1-rc4 2025-09-07T07:28:19.0872210Z * [new tag] v2.1.0 -> v2.1.0 2025-09-07T07:28:19.0872978Z * [new tag] v2.1.0-rc1 -> v2.1.0-rc1 2025-09-07T07:28:19.0873704Z * [new tag] v2.1.0-rc2 -> v2.1.0-rc2 2025-09-07T07:28:19.0874498Z * [new tag] v2.1.0-rc3 -> v2.1.0-rc3 2025-09-07T07:28:19.0875233Z * [new tag] v2.1.0-rc4 -> v2.1.0-rc4 2025-09-07T07:28:19.0876014Z * [new tag] v2.1.0-rc5 -> v2.1.0-rc5 2025-09-07T07:28:19.0876731Z * [new tag] v2.1.0-rc6 -> v2.1.0-rc6 2025-09-07T07:28:19.0877310Z * [new tag] v2.1.1 -> v2.1.1 2025-09-07T07:28:19.0878128Z * [new tag] v2.1.1-rc1 -> v2.1.1-rc1 2025-09-07T07:28:19.0878843Z * [new tag] v2.1.1-rc2 -> v2.1.1-rc2 2025-09-07T07:28:19.0879651Z * [new tag] v2.1.1-rc3 -> v2.1.1-rc3 2025-09-07T07:28:19.0880391Z * [new tag] v2.1.1-rc4 -> v2.1.1-rc4 2025-09-07T07:28:19.0881136Z * [new tag] v2.1.1-rc5 -> v2.1.1-rc5 2025-09-07T07:28:19.0881664Z * [new tag] v2.1.1-rc6 -> v2.1.1-rc6 2025-09-07T07:28:19.0882353Z * [new tag] v2.1.2 -> v2.1.2 2025-09-07T07:28:19.0883080Z * [new tag] v2.1.2-rc1 -> v2.1.2-rc1 2025-09-07T07:28:19.0883850Z * [new tag] v2.1.2-rc2 -> v2.1.2-rc2 2025-09-07T07:28:19.0884427Z * [new tag] v2.1.2-rc3 -> v2.1.2-rc3 2025-09-07T07:28:19.0885225Z * [new tag] v2.2.0 -> v2.2.0 2025-09-07T07:28:19.0885972Z * [new tag] v2.2.0-rc1 -> v2.2.0-rc1 2025-09-07T07:28:19.0886621Z * [new tag] v2.2.0-rc2 -> v2.2.0-rc2 2025-09-07T07:28:19.0887332Z * [new tag] v2.2.0-rc3 -> v2.2.0-rc3 2025-09-07T07:28:19.0887997Z * [new tag] v2.2.0-rc4 -> v2.2.0-rc4 2025-09-07T07:28:19.0888768Z * [new tag] v2.2.0-rc5 -> v2.2.0-rc5 2025-09-07T07:28:19.0889455Z * [new tag] v2.2.0-rc6 -> v2.2.0-rc6 2025-09-07T07:28:19.0889987Z * [new tag] v2.2.0-rc7 -> v2.2.0-rc7 2025-09-07T07:28:19.0890556Z * [new tag] v2.2.0-rc8 -> v2.2.0-rc8 2025-09-07T07:28:19.0891309Z * [new tag] v2.2.1 -> v2.2.1 2025-09-07T07:28:19.0892112Z * [new tag] v2.2.1-rc1 -> v2.2.1-rc1 2025-09-07T07:28:19.0892631Z * [new tag] v2.2.1-rc2 -> v2.2.1-rc2 2025-09-07T07:28:19.0893171Z * [new tag] v2.2.1-rc3 -> v2.2.1-rc3 2025-09-07T07:28:19.0893802Z * [new tag] v2.2.2 -> v2.2.2 2025-09-07T07:28:19.0894607Z * [new tag] v2.2.2-rc1 -> v2.2.2-rc1 2025-09-07T07:28:19.0895102Z * [new tag] v2.2.2-rc2 -> v2.2.2-rc2 2025-09-07T07:28:19.0895670Z * [new tag] v2.2.2-rc3 -> v2.2.2-rc3 2025-09-07T07:28:19.0896396Z * [new tag] v2.3.0 -> v2.3.0 2025-09-07T07:28:19.0897083Z * [new tag] v2.3.0-rc1 -> v2.3.0-rc1 2025-09-07T07:28:19.0897886Z * [new tag] v2.3.0-rc10 -> v2.3.0-rc10 2025-09-07T07:28:19.0898661Z * [new tag] v2.3.0-rc11 -> v2.3.0-rc11 2025-09-07T07:28:19.0899139Z * [new tag] v2.3.0-rc12 -> v2.3.0-rc12 2025-09-07T07:28:19.0900002Z * [new tag] v2.3.0-rc2 -> v2.3.0-rc2 2025-09-07T07:28:19.0901210Z * [new tag] v2.3.0-rc3 -> v2.3.0-rc3 2025-09-07T07:28:19.0901943Z * [new tag] v2.3.0-rc4 -> v2.3.0-rc4 2025-09-07T07:28:19.0902765Z * [new tag] v2.3.0-rc5 -> v2.3.0-rc5 2025-09-07T07:28:19.0903280Z * [new tag] v2.3.0-rc6 -> v2.3.0-rc6 2025-09-07T07:28:19.0904034Z * [new tag] v2.3.0-rc7 -> v2.3.0-rc7 2025-09-07T07:28:19.0904730Z * [new tag] v2.3.0-rc8 -> v2.3.0-rc8 2025-09-07T07:28:19.0905243Z * [new tag] v2.3.0-rc9 -> v2.3.0-rc9 2025-09-07T07:28:19.0905904Z * [new tag] v2.3.1 -> v2.3.1 2025-09-07T07:28:19.0906527Z * [new tag] v2.3.1-rc1 -> v2.3.1-rc1 2025-09-07T07:28:19.0907267Z * [new tag] v2.3.1-rc2 -> v2.3.1-rc2 2025-09-07T07:28:19.0907979Z * [new tag] v2.3.1-rc3 -> v2.3.1-rc3 2025-09-07T07:28:19.0908767Z * [new tag] v2.4.0 -> v2.4.0 2025-09-07T07:28:19.0909495Z * [new tag] v2.4.0-rc1 -> v2.4.0-rc1 2025-09-07T07:28:19.0910215Z * [new tag] v2.4.0-rc2 -> v2.4.0-rc2 2025-09-07T07:28:19.0910912Z * [new tag] v2.4.0-rc3 -> v2.4.0-rc3 2025-09-07T07:28:19.0911659Z * [new tag] v2.4.0-rc4 -> v2.4.0-rc4 2025-09-07T07:28:19.0912385Z * [new tag] v2.4.0-rc5 -> v2.4.0-rc5 2025-09-07T07:28:19.0913196Z * [new tag] v2.4.0-rc6 -> v2.4.0-rc6 2025-09-07T07:28:19.0913976Z * [new tag] v2.4.0-rc7 -> v2.4.0-rc7 2025-09-07T07:28:19.0914655Z * [new tag] v2.4.0-rc8 -> v2.4.0-rc8 2025-09-07T07:28:19.0915546Z * [new tag] v2.4.0-rc9 -> v2.4.0-rc9 2025-09-07T07:28:19.0916082Z * [new tag] v2.4.1 -> v2.4.1 2025-09-07T07:28:19.0916903Z * [new tag] v2.4.1-rc1 -> v2.4.1-rc1 2025-09-07T07:28:19.0917658Z * [new tag] v2.4.1-rc2 -> v2.4.1-rc2 2025-09-07T07:28:19.0918427Z * [new tag] v2.4.1-rc3 -> v2.4.1-rc3 2025-09-07T07:28:19.0919179Z * [new tag] v2.5.0 -> v2.5.0 2025-09-07T07:28:19.0919873Z * [new tag] v2.5.0-rc1 -> v2.5.0-rc1 2025-09-07T07:28:19.0920453Z * [new tag] v2.5.0-rc10 -> v2.5.0-rc10 2025-09-07T07:28:19.0921156Z * [new tag] v2.5.0-rc2 -> v2.5.0-rc2 2025-09-07T07:28:19.0921908Z * [new tag] v2.5.0-rc3 -> v2.5.0-rc3 2025-09-07T07:28:19.0922630Z * [new tag] v2.5.0-rc4 -> v2.5.0-rc4 2025-09-07T07:28:19.0923421Z * [new tag] v2.5.0-rc5 -> v2.5.0-rc5 2025-09-07T07:28:19.0924199Z * [new tag] v2.5.0-rc6 -> v2.5.0-rc6 2025-09-07T07:28:19.0924976Z * [new tag] v2.5.0-rc7 -> v2.5.0-rc7 2025-09-07T07:28:19.0925795Z * [new tag] v2.5.0-rc8 -> v2.5.0-rc8 2025-09-07T07:28:19.0926495Z * [new tag] v2.5.0-rc9 -> v2.5.0-rc9 2025-09-07T07:28:19.0927005Z * [new tag] v2.5.1 -> v2.5.1 2025-09-07T07:28:19.0927645Z * [new tag] v2.5.1-rc1 -> v2.5.1-rc1 2025-09-07T07:28:19.0928140Z * [new tag] v2.6.0 -> v2.6.0 2025-09-07T07:28:19.0928966Z * [new tag] v2.6.0-rc1 -> v2.6.0-rc1 2025-09-07T07:28:19.0929755Z * [new tag] v2.6.0-rc2 -> v2.6.0-rc2 2025-09-07T07:28:19.0930456Z * [new tag] v2.6.0-rc3 -> v2.6.0-rc3 2025-09-07T07:28:19.0931162Z * [new tag] v2.6.0-rc4 -> v2.6.0-rc4 2025-09-07T07:28:19.0932105Z * [new tag] v2.6.0-rc5 -> v2.6.0-rc5 2025-09-07T07:28:19.0932972Z * [new tag] v2.6.0-rc6 -> v2.6.0-rc6 2025-09-07T07:28:19.0933742Z * [new tag] v2.6.0-rc7 -> v2.6.0-rc7 2025-09-07T07:28:19.0934605Z * [new tag] v2.6.0-rc8 -> v2.6.0-rc8 2025-09-07T07:28:19.0935358Z * [new tag] v2.6.0-rc9 -> v2.6.0-rc9 2025-09-07T07:28:19.0936332Z * [new tag] v2.7.0 -> v2.7.0 2025-09-07T07:28:19.0937015Z * [new tag] v2.7.0-rc1 -> v2.7.0-rc1 2025-09-07T07:28:19.0937586Z * [new tag] v2.7.0-rc10 -> v2.7.0-rc10 2025-09-07T07:28:19.0938374Z * [new tag] v2.7.0-rc2 -> v2.7.0-rc2 2025-09-07T07:28:19.0939128Z * [new tag] v2.7.0-rc3 -> v2.7.0-rc3 2025-09-07T07:28:19.0939901Z * [new tag] v2.7.0-rc4 -> v2.7.0-rc4 2025-09-07T07:28:19.0940646Z * [new tag] v2.7.0-rc5 -> v2.7.0-rc5 2025-09-07T07:28:19.0941375Z * [new tag] v2.7.0-rc6 -> v2.7.0-rc6 2025-09-07T07:28:19.0942150Z * [new tag] v2.7.0-rc7 -> v2.7.0-rc7 2025-09-07T07:28:19.0942949Z * [new tag] v2.7.0-rc8 -> v2.7.0-rc8 2025-09-07T07:28:19.0943751Z * [new tag] v2.7.0-rc9 -> v2.7.0-rc9 2025-09-07T07:28:19.0944257Z * [new tag] v2.7.1 -> v2.7.1 2025-09-07T07:28:19.0945046Z * [new tag] v2.7.1-rc1 -> v2.7.1-rc1 2025-09-07T07:28:19.0945843Z * [new tag] v2.7.1-rc2 -> v2.7.1-rc2 2025-09-07T07:28:19.0946727Z * [new tag] v2.7.1-rc3 -> v2.7.1-rc3 2025-09-07T07:28:19.0947618Z * [new tag] v2.7.1-rc4 -> v2.7.1-rc4 2025-09-07T07:28:19.0948496Z * [new tag] v2.7.1-rc5 -> v2.7.1-rc5 2025-09-07T07:28:19.0949125Z * [new tag] v2.8.0 -> v2.8.0 2025-09-07T07:28:19.0949898Z * [new tag] v2.8.0-rc1 -> v2.8.0-rc1 2025-09-07T07:28:19.0950676Z * [new tag] v2.8.0-rc2 -> v2.8.0-rc2 2025-09-07T07:28:19.0951449Z * [new tag] v2.8.0-rc3 -> v2.8.0-rc3 2025-09-07T07:28:19.0952261Z * [new tag] v2.8.0-rc4 -> v2.8.0-rc4 2025-09-07T07:28:19.0953065Z * [new tag] v2.8.0-rc5 -> v2.8.0-rc5 2025-09-07T07:28:19.0953827Z * [new tag] v2.8.0-rc6 -> v2.8.0-rc6 2025-09-07T07:28:19.0954596Z * [new tag] v2.8.0-rc7 -> v2.8.0-rc7 2025-09-07T07:28:19.0955286Z * [new tag] v2.8.0-rc8 -> v2.8.0-rc8 2025-09-07T07:28:19.0956039Z * [new tag] whc_flight_1 -> whc_flight_1 2025-09-07T07:28:19.0956864Z * [new tag] whc_flight_2 -> whc_flight_2 2025-09-07T07:28:19.0957534Z * [new tag] whc_flight_4 -> whc_flight_4 2025-09-07T07:28:19.1531459Z [command]/usr/bin/git rev-parse --verify --quiet 93fb23d6fae7c4e82c4239a1033e522088742634^{object} 2025-09-07T07:28:19.1556279Z 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:28:19.1560228Z ##[endgroup] 2025-09-07T07:28:19.1560485Z ##[group]Determining the checkout info 2025-09-07T07:28:19.1561449Z ##[endgroup] 2025-09-07T07:28:19.1565157Z [command]/usr/bin/git sparse-checkout disable 2025-09-07T07:28:19.1598050Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-09-07T07:28:19.1625988Z ##[group]Checking out the ref 2025-09-07T07:28:19.1629611Z [command]/usr/bin/git checkout --progress --force 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:28:20.1900635Z Updating files: 90% (17564/19405) 2025-09-07T07:28:20.1992164Z Updating files: 91% (17659/19405) 2025-09-07T07:28:20.2106869Z Updating files: 92% (17853/19405) 2025-09-07T07:28:20.2260320Z Updating files: 93% (18047/19405) 2025-09-07T07:28:20.2427246Z Updating files: 94% (18241/19405) 2025-09-07T07:28:20.2555053Z Updating files: 95% (18435/19405) 2025-09-07T07:28:20.2684049Z Updating files: 96% (18629/19405) 2025-09-07T07:28:20.2834844Z Updating files: 97% (18823/19405) 2025-09-07T07:28:20.3057370Z Updating files: 98% (19017/19405) 2025-09-07T07:28:20.3186210Z Updating files: 99% (19211/19405) 2025-09-07T07:28:20.3186472Z Updating files: 100% (19405/19405) 2025-09-07T07:28:20.3186727Z Updating files: 100% (19405/19405), done. 2025-09-07T07:28:20.3455072Z Note: switching to '93fb23d6fae7c4e82c4239a1033e522088742634'. 2025-09-07T07:28:20.3455340Z 2025-09-07T07:28:20.3455529Z You are in 'detached HEAD' state. You can look around, make experimental 2025-09-07T07:28:20.3456014Z changes and commit them, and you can discard any commits you make in this 2025-09-07T07:28:20.3456487Z state without impacting any branches by switching back to a branch. 2025-09-07T07:28:20.3456774Z 2025-09-07T07:28:20.3456961Z If you want to create a new branch to retain commits you create, you may 2025-09-07T07:28:20.3457393Z do so (now or later) by using -c with the switch command. Example: 2025-09-07T07:28:20.3457647Z 2025-09-07T07:28:20.3457742Z git switch -c 2025-09-07T07:28:20.3457929Z 2025-09-07T07:28:20.3458020Z Or undo this operation with: 2025-09-07T07:28:20.3458177Z 2025-09-07T07:28:20.3458255Z git switch - 2025-09-07T07:28:20.3458368Z 2025-09-07T07:28:20.3458581Z Turn off this advice by setting config variable advice.detachedHead to false 2025-09-07T07:28:20.3458884Z 2025-09-07T07:28:20.3459042Z HEAD is now at 93fb23d6fae Build vLLM nightly wheels (#162000) 2025-09-07T07:28:20.3550082Z ##[endgroup] 2025-09-07T07:28:20.3550451Z ##[group]Setting up auth for fetching submodules 2025-09-07T07:28:20.3555970Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-09-07T07:28:20.3600977Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-09-07T07:28:20.3628350Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-09-07T07:28:20.3654453Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-09-07T07:28:20.3678343Z ##[endgroup] 2025-09-07T07:28:20.3678699Z ##[group]Fetching submodules 2025-09-07T07:28:20.3681468Z [command]/usr/bin/git submodule sync --recursive 2025-09-07T07:28:20.3968976Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --recursive 2025-09-07T07:28:20.4592655Z Submodule 'android/libs/fbjni' (https://github.com/facebookincubator/fbjni.git) registered for path 'android/libs/fbjni' 2025-09-07T07:28:20.4594821Z Submodule 'third_party/NNPACK_deps/FP16' (https://github.com/Maratyszcza/FP16.git) registered for path 'third_party/FP16' 2025-09-07T07:28:20.4597394Z Submodule 'third_party/NNPACK_deps/FXdiv' (https://github.com/Maratyszcza/FXdiv.git) registered for path 'third_party/FXdiv' 2025-09-07T07:28:20.4600019Z Submodule 'third_party/NNPACK' (https://github.com/Maratyszcza/NNPACK.git) registered for path 'third_party/NNPACK' 2025-09-07T07:28:20.4602722Z Submodule 'third_party/NVTX' (https://github.com/NVIDIA/NVTX.git) registered for path 'third_party/NVTX' 2025-09-07T07:28:20.4617331Z Submodule 'third_party/VulkanMemoryAllocator' (https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator.git) registered for path 'third_party/VulkanMemoryAllocator' 2025-09-07T07:28:20.4619871Z Submodule 'third_party/XNNPACK' (https://github.com/google/XNNPACK.git) registered for path 'third_party/XNNPACK' 2025-09-07T07:28:20.4622717Z Submodule 'third_party/aiter' (https://github.com/ROCm/aiter.git) registered for path 'third_party/aiter' 2025-09-07T07:28:20.4625590Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/benchmark' 2025-09-07T07:28:20.4628626Z Submodule 'third_party/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/composable_kernel' 2025-09-07T07:28:20.4631521Z Submodule 'third_party/cpp-httplib' (https://github.com/yhirose/cpp-httplib.git) registered for path 'third_party/cpp-httplib' 2025-09-07T07:28:20.4646689Z Submodule 'third_party/cpuinfo' (https://github.com/pytorch/cpuinfo.git) registered for path 'third_party/cpuinfo' 2025-09-07T07:28:20.4649709Z Submodule 'third_party/cudnn_frontend' (https://github.com/NVIDIA/cudnn-frontend.git) registered for path 'third_party/cudnn_frontend' 2025-09-07T07:28:20.4652802Z Submodule 'third_party/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/cutlass' 2025-09-07T07:28:20.4656016Z Submodule 'third_party/fbgemm' (https://github.com/pytorch/fbgemm) registered for path 'third_party/fbgemm' 2025-09-07T07:28:20.4661185Z Submodule 'third_party/flash-attention' (https://github.com/Dao-AILab/flash-attention.git) registered for path 'third_party/flash-attention' 2025-09-07T07:28:20.4676859Z Submodule 'third_party/flatbuffers' (https://github.com/google/flatbuffers.git) registered for path 'third_party/flatbuffers' 2025-09-07T07:28:20.4680305Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/fmt' 2025-09-07T07:28:20.4683951Z Submodule 'third_party/gemmlowp/gemmlowp' (https://github.com/google/gemmlowp.git) registered for path 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:28:20.4687394Z Submodule 'third_party/gloo' (https://github.com/pytorch/gloo) registered for path 'third_party/gloo' 2025-09-07T07:28:20.4691107Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/googletest' 2025-09-07T07:28:20.4694728Z Submodule 'third_party/ideep' (https://github.com/intel/ideep) registered for path 'third_party/ideep' 2025-09-07T07:28:20.4709317Z Submodule 'third_party/ittapi' (https://github.com/intel/ittapi.git) registered for path 'third_party/ittapi' 2025-09-07T07:28:20.4713084Z Submodule 'third_party/kineto' (https://github.com/pytorch/kineto) registered for path 'third_party/kineto' 2025-09-07T07:28:20.4716991Z Submodule 'third_party/kleidiai' (https://github.com/ARM-software/kleidiai.git) registered for path 'third_party/kleidiai' 2025-09-07T07:28:20.4720889Z Submodule 'third_party/mimalloc' (https://github.com/microsoft/mimalloc.git) registered for path 'third_party/mimalloc' 2025-09-07T07:28:20.4724852Z Submodule 'third_party/nlohmann' (https://github.com/nlohmann/json.git) registered for path 'third_party/nlohmann' 2025-09-07T07:28:20.4739924Z Submodule 'third_party/onnx' (https://github.com/onnx/onnx.git) registered for path 'third_party/onnx' 2025-09-07T07:28:20.4744313Z Submodule 'third_party/opentelemetry-cpp' (https://github.com/open-telemetry/opentelemetry-cpp.git) registered for path 'third_party/opentelemetry-cpp' 2025-09-07T07:28:20.4748314Z Submodule 'third_party/pocketfft' (https://github.com/mreineck/pocketfft) registered for path 'third_party/pocketfft' 2025-09-07T07:28:20.4752605Z Submodule 'third_party/protobuf' (https://github.com/protocolbuffers/protobuf.git) registered for path 'third_party/protobuf' 2025-09-07T07:28:20.4756907Z Submodule 'third_party/NNPACK_deps/psimd' (https://github.com/Maratyszcza/psimd.git) registered for path 'third_party/psimd' 2025-09-07T07:28:20.4761351Z Submodule 'third_party/NNPACK_deps/pthreadpool' (https://github.com/Maratyszcza/pthreadpool.git) registered for path 'third_party/pthreadpool' 2025-09-07T07:28:20.4778450Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/pybind11' 2025-09-07T07:28:20.4783089Z Submodule 'third_party/python-peachpy' (https://github.com/malfet/PeachPy.git) registered for path 'third_party/python-peachpy' 2025-09-07T07:28:20.4787626Z Submodule 'third_party/sleef' (https://github.com/shibatch/sleef) registered for path 'third_party/sleef' 2025-09-07T07:28:20.4792287Z Submodule 'third_party/tensorpipe' (https://github.com/pytorch/tensorpipe.git) registered for path 'third_party/tensorpipe' 2025-09-07T07:28:20.4821107Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/android/libs/fbjni'... 2025-09-07T07:28:20.7164443Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FXdiv'... 2025-09-07T07:28:20.7165453Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FP16'... 2025-09-07T07:28:20.7166136Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/psimd'... 2025-09-07T07:28:20.7166746Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NNPACK'... 2025-09-07T07:28:20.7180753Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NVTX'... 2025-09-07T07:28:20.8577772Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pthreadpool'... 2025-09-07T07:28:20.8578467Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pocketfft'... 2025-09-07T07:28:20.8579553Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/python-peachpy'... 2025-09-07T07:28:20.8580235Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep'... 2025-09-07T07:28:20.8580949Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gemmlowp/gemmlowp'... 2025-09-07T07:28:20.8581607Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gloo'... 2025-09-07T07:28:20.8582336Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/benchmark'... 2025-09-07T07:28:20.8642748Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/VulkanMemoryAllocator'... 2025-09-07T07:28:21.5814881Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe'... 2025-09-07T07:28:21.5815560Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ittapi'... 2025-09-07T07:28:21.5816189Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kleidiai'... 2025-09-07T07:28:21.5816860Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpp-httplib'... 2025-09-07T07:28:21.5817523Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention'... 2025-09-07T07:28:21.5818185Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpuinfo'... 2025-09-07T07:28:21.5818802Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/googletest'... 2025-09-07T07:28:21.5819424Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/mimalloc'... 2025-09-07T07:28:21.5820074Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/sleef'... 2025-09-07T07:28:21.5820679Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pybind11'... 2025-09-07T07:28:21.5821277Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fmt'... 2025-09-07T07:28:21.6580785Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/XNNPACK'... 2025-09-07T07:28:30.1106601Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cudnn_frontend'... 2025-09-07T07:28:30.1107341Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flatbuffers'... 2025-09-07T07:28:30.1107982Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto'... 2025-09-07T07:28:30.1108587Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm'... 2025-09-07T07:28:30.1109192Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cutlass'... 2025-09-07T07:28:30.1109788Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx'... 2025-09-07T07:28:30.1110423Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/composable_kernel'... 2025-09-07T07:28:30.1111063Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter'... 2025-09-07T07:28:30.1111714Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp'... 2025-09-07T07:28:30.1112366Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/nlohmann'... 2025-09-07T07:28:30.1113326Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf'... 2025-09-07T07:28:30.1254070Z Submodule path 'android/libs/fbjni': checked out '7e1e1fe3858c63c251c637ae41a20de425dde96f' 2025-09-07T07:28:30.1370621Z Submodule path 'third_party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2025-09-07T07:28:30.1464592Z Submodule path 'third_party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2025-09-07T07:28:30.1689417Z Submodule path 'third_party/NNPACK': checked out 'c07e3a0400713d546e0dea2d5466dd22ea389c73' 2025-09-07T07:28:30.2390808Z Submodule path 'third_party/NVTX': checked out '2942f167cc30c5e3a44a2aecd5b0d9c07ff61a07' 2025-09-07T07:28:30.2844655Z Submodule path 'third_party/VulkanMemoryAllocator': checked out '1d8f600fd424278486eade7ed3e877c99f0846b1' 2025-09-07T07:28:30.8937032Z Submodule path 'third_party/XNNPACK': checked out '51a0103656eff6fc9bfd39a4597923c4b542c883' 2025-09-07T07:28:31.0288235Z Submodule path 'third_party/aiter': checked out '01aae101b9e5e94d6c16a9514c9fb8df99c93150' 2025-09-07T07:28:31.0307726Z Submodule '3rdparty/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:28:31.0331951Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter/3rdparty/composable_kernel'... 2025-09-07T07:28:34.1460177Z Submodule path 'third_party/aiter/3rdparty/composable_kernel': checked out 'cffe8fa2a442ac8e80dd236a1a5d24fe3d7e0cbf' 2025-09-07T07:28:34.1671895Z Submodule path 'third_party/benchmark': checked out '299e5928955cc62af9968370293b916f5130916f' 2025-09-07T07:28:34.4428606Z Submodule path 'third_party/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-09-07T07:28:34.4879361Z Submodule path 'third_party/cpp-httplib': checked out '89c932f313c6437c38f2982869beacc89c2f2246' 2025-09-07T07:28:34.5748902Z Submodule path 'third_party/cpuinfo': checked out '5e3d2445e6a84d9599bee2bf78edbb4d80865e1d' 2025-09-07T07:28:34.6145639Z Submodule path 'third_party/cudnn_frontend': checked out 'f937055efc6d414d11f4c6577e3977fe74f35fb6' 2025-09-07T07:28:35.1653492Z Submodule path 'third_party/cutlass': checked out 'e51efbfe18fe4f4cbb66ab814c55bf4aa0185491' 2025-09-07T07:28:35.2941244Z Submodule path 'third_party/fbgemm': checked out '4b39c551efe15e6bbade20565b0ceb2d8ce3352d' 2025-09-07T07:28:35.2960945Z Submodule 'external/asmjit' (https://github.com/asmjit/asmjit.git) registered for path 'third_party/fbgemm/external/asmjit' 2025-09-07T07:28:35.2963237Z Submodule 'external/composable_kernel' (https://github.com/jwfromm/composable_kernel.git) registered for path 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:28:35.2965313Z Submodule 'external/cpuinfo' (https://github.com/pytorch/cpuinfo) registered for path 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:28:35.2967582Z Submodule 'external/cutlass' (https://github.com/jwfromm/cutlass) registered for path 'third_party/fbgemm/external/cutlass' 2025-09-07T07:28:35.2969841Z Submodule 'external/googletest' (https://github.com/google/googletest) registered for path 'third_party/fbgemm/external/googletest' 2025-09-07T07:28:35.2972243Z Submodule 'external/hipify_torch' (https://github.com/ROCmSoftwarePlatform/hipify_torch.git) registered for path 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:28:35.2974471Z Submodule 'external/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/fbgemm/external/json' 2025-09-07T07:28:35.3000430Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/asmjit'... 2025-09-07T07:28:36.3009736Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/hipify_torch'... 2025-09-07T07:28:36.3010604Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cpuinfo'... 2025-09-07T07:28:36.3011357Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/googletest'... 2025-09-07T07:28:36.3012441Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/composable_kernel'... 2025-09-07T07:28:36.4009721Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cutlass'... 2025-09-07T07:28:37.0354109Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/json'... 2025-09-07T07:28:41.3914347Z Submodule path 'third_party/fbgemm/external/asmjit': checked out 'a3199e8857792cd10b7589ff5d58343d2c9008ea' 2025-09-07T07:28:41.6115956Z Submodule path 'third_party/fbgemm/external/composable_kernel': checked out 'b1281b8b08d973a7064f864f47eeb30f3e2596e9' 2025-09-07T07:28:41.7004033Z Submodule path 'third_party/fbgemm/external/cpuinfo': checked out '6543fec09b2f04ac4a666882998b534afc9c1349' 2025-09-07T07:28:42.2513675Z Submodule path 'third_party/fbgemm/external/cutlass': checked out '311f3c8e51dc0eb56310cfc6980bf63d0fbd7917' 2025-09-07T07:28:42.2908275Z Submodule path 'third_party/fbgemm/external/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-09-07T07:28:42.3018771Z Submodule path 'third_party/fbgemm/external/hipify_torch': checked out '63b6a7b541fa7f08f8475ca7d74054db36ff2691' 2025-09-07T07:28:42.4064086Z Submodule path 'third_party/fbgemm/external/json': checked out '9cca280a4d0ccf0c08f47a99aa71d1b0e52f8d03' 2025-09-07T07:28:42.4696083Z Submodule path 'third_party/flash-attention': checked out '979702c87a8713a8e0a5e9fee122b90d2ef13be5' 2025-09-07T07:28:42.4713441Z Submodule 'csrc/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:28:42.4715386Z Submodule 'csrc/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:28:42.4739997Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/composable_kernel'... 2025-09-07T07:28:45.3076530Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/cutlass'... 2025-09-07T07:28:45.5115532Z Submodule path 'third_party/flash-attention/csrc/composable_kernel': checked out '888317e698e9803c62bd38568abc9e05d7709f33' 2025-09-07T07:28:46.0058814Z Submodule path 'third_party/flash-attention/csrc/cutlass': checked out 'c506e16788cb08416a4a57e11a9067beeee29420' 2025-09-07T07:28:46.1203221Z Submodule path 'third_party/flatbuffers': checked out 'a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757' 2025-09-07T07:28:46.1495969Z Submodule path 'third_party/fmt': checked out '40626af88bd7df9a5fb80be7b25ac85b122d6c21' 2025-09-07T07:28:46.1832903Z Submodule path 'third_party/gemmlowp/gemmlowp': checked out '3fb5c176c17c765a3492cd2f0321b0dab712f350' 2025-09-07T07:28:46.2062990Z Submodule path 'third_party/gloo': checked out 'c7b7b022c124d9643957d9bd55f57ac59fce8fa2' 2025-09-07T07:28:46.2457923Z Submodule path 'third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-09-07T07:28:46.2583627Z Submodule path 'third_party/ideep': checked out '719d8e6cd7f7a0e01b155657526d693acf97c2b3' 2025-09-07T07:28:46.2600304Z Submodule 'mkl-dnn' (https://github.com/intel/mkl-dnn.git) registered for path 'third_party/ideep/mkl-dnn' 2025-09-07T07:28:46.2624000Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep/mkl-dnn'... 2025-09-07T07:28:56.9335811Z Submodule path 'third_party/ideep/mkl-dnn': checked out '8d263e693366ef8db40acc569cc7d8edf644556d' 2025-09-07T07:28:56.9522067Z Submodule path 'third_party/ittapi': checked out 'dec1d23ca65ab069d225dfe40dea14f455170959' 2025-09-07T07:28:57.0494685Z Submodule path 'third_party/kineto': checked out '5e7501833f1021ce6f618572d3baf657b6319658' 2025-09-07T07:28:57.0512521Z Submodule 'libkineto/third_party/dynolog' (https://github.com/facebookincubator/dynolog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:28:57.0514649Z Submodule 'libkineto/third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:28:57.0516636Z Submodule 'libkineto/third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:28:57.0541581Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog'... 2025-09-07T07:28:57.6263354Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/fmt'... 2025-09-07T07:28:57.8873097Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/googletest'... 2025-09-07T07:28:57.9609491Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog': checked out '7d04a0053a845370ae06ce317a22a48e9edcc74e' 2025-09-07T07:28:57.9627076Z Submodule 'third_party/DCGM' (https://github.com/NVIDIA/DCGM.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:28:57.9629260Z Submodule 'third_party/cpr' (https://github.com/libcpr/cpr.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:28:57.9631438Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:28:57.9633659Z Submodule 'third_party/gflags' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:28:57.9635929Z Submodule 'third_party/glog' (https://github.com/google/glog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:28:57.9638426Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:28:57.9640812Z Submodule 'third_party/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:28:57.9643280Z Submodule 'third_party/pfs' (https://github.com/dtrugman/pfs.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:28:57.9669994Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM'... 2025-09-07T07:28:59.1029375Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/pfs'... 2025-09-07T07:28:59.1039228Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags'... 2025-09-07T07:28:59.1040589Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/cpr'... 2025-09-07T07:28:59.1041581Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/glog'... 2025-09-07T07:28:59.1042599Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/googletest'... 2025-09-07T07:28:59.1043596Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/fmt'... 2025-09-07T07:28:59.2029443Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/json'... 2025-09-07T07:29:03.9233631Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM': checked out 'ffde4e54bc7249a6039a5e6b45b395141e1217f9' 2025-09-07T07:29:03.9397677Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr': checked out '871ed52d350214a034f6ef8a3b8f51c5ce1bd400' 2025-09-07T07:29:03.9714658Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt': checked out 'cd4af11efc9c622896a3e4cb599fa28668ca3d05' 2025-09-07T07:29:03.9841724Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags': checked out 'e171aa2d15ed9eb17054558e0b3a6a413bb01067' 2025-09-07T07:29:03.9857114Z Submodule 'doc' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:29:03.9882038Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc'... 2025-09-07T07:29:04.2381500Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc': checked out '8411df715cf522606e3b1aca386ddfc0b63d34b4' 2025-09-07T07:29:04.2549668Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog': checked out 'b33e3bad4c46c8a6345525fd822af355e5ef9446' 2025-09-07T07:29:04.2927879Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest': checked out '58d77fa8070e8cec2dc1ed015d66b454c8d78850' 2025-09-07T07:29:04.3821938Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json': checked out '4f8fba14066156b73f1189a2b8bd568bde5284c5' 2025-09-07T07:29:04.3975624Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs': checked out 'f68a2fa8ea36c783bdd760371411fcb495aa3150' 2025-09-07T07:29:04.4301707Z Submodule path 'third_party/kineto/libkineto/third_party/fmt': checked out '0041a40c1350ba702d475b9c4ad62da77caea164' 2025-09-07T07:29:04.4817004Z Submodule path 'third_party/kineto/libkineto/third_party/googletest': checked out '7aca84427f224eeed3144123d5230d5871e93347' 2025-09-07T07:29:04.5245227Z Submodule path 'third_party/kleidiai': checked out 'cca02c2f69dd18e1f12647c1c0bdc8cf90e680c7' 2025-09-07T07:29:04.5593197Z Submodule path 'third_party/mimalloc': checked out 'fbd8b99c2b828428947d70fdc046bb55609be93e' 2025-09-07T07:29:04.6730669Z Submodule path 'third_party/nlohmann': checked out '55f93686c01528224f448c19128836e7df245f72' 2025-09-07T07:29:05.0209914Z Submodule path 'third_party/onnx': checked out 'e709452ef2bbc1d113faf678c24e6d3467696e83' 2025-09-07T07:29:05.0241760Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/onnx/third_party/pybind11' 2025-09-07T07:29:05.0266650Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx/third_party/pybind11'... 2025-09-07T07:29:06.0323562Z Submodule path 'third_party/onnx/third_party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-09-07T07:29:06.0935447Z Submodule path 'third_party/opentelemetry-cpp': checked out 'a799f4aed9c94b765dcdaabaeab7d5e7e2310878' 2025-09-07T07:29:06.0954664Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark) registered for path 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:29:06.0956832Z Submodule 'third_party/googletest' (https://github.com/google/googletest) registered for path 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:29:06.0958956Z Submodule 'third_party/ms-gsl' (https://github.com/microsoft/GSL) registered for path 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:29:06.0961205Z Submodule 'third_party/nlohmann-json' (https://github.com/nlohmann/json) registered for path 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:29:06.0963627Z Submodule 'third_party/opentelemetry-proto' (https://github.com/open-telemetry/opentelemetry-proto) registered for path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:29:06.0965886Z Submodule 'third_party/opentracing-cpp' (https://github.com/opentracing/opentracing-cpp.git) registered for path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:29:06.0968451Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:29:06.0970519Z Submodule 'tools/vcpkg' (https://github.com/Microsoft/vcpkg) registered for path 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:29:06.0996504Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/benchmark'... 2025-09-07T07:29:06.5530180Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentracing-cpp'... 2025-09-07T07:29:06.5531186Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentelemetry-proto'... 2025-09-07T07:29:06.5532093Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/ms-gsl'... 2025-09-07T07:29:06.5533001Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp'... 2025-09-07T07:29:06.6531238Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/googletest'... 2025-09-07T07:29:07.2287788Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/nlohmann-json'... 2025-09-07T07:29:12.5882366Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/tools/vcpkg'... 2025-09-07T07:29:12.6055987Z Submodule path 'third_party/opentelemetry-cpp/third_party/benchmark': checked out 'd572f4777349d43653b21d6c2fc63020ab326db2' 2025-09-07T07:29:12.6413197Z Submodule path 'third_party/opentelemetry-cpp/third_party/googletest': checked out 'b796f7d44681514f58a683a3a71ff17c94edb0c1' 2025-09-07T07:29:12.6573919Z Submodule path 'third_party/opentelemetry-cpp/third_party/ms-gsl': checked out '6f4529395c5b7c2d661812257cd6780c67e54afa' 2025-09-07T07:29:12.7631416Z Submodule path 'third_party/opentelemetry-cpp/third_party/nlohmann-json': checked out 'bc889afb4c5bf1c0d8ee29ef35eaaf4c8bef8a5d' 2025-09-07T07:29:12.7761225Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto': checked out '4ca4f0335c63cda7ab31ea7ed70d6553aee14dce' 2025-09-07T07:29:12.7900179Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp': checked out '06b57f48ded1fa3bdd3d4346f6ef29e40e08eaf5' 2025-09-07T07:29:12.8046206Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp': checked out 'c9ffcdda9086ffd9e1283ea7a0276d831f3c8a8d' 2025-09-07T07:29:12.8061899Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:29:12.8064150Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:29:12.8089003Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-09-07T07:29:14.5927538Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest'... 2025-09-07T07:29:14.8033061Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'eefb26f82b233268fc98577d265352720d477ba4' 2025-09-07T07:29:14.8426349Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-09-07T07:29:15.2434800Z Submodule path 'third_party/opentelemetry-cpp/tools/vcpkg': checked out '8eb57355a4ffb410a2e94c07b4dca2dffbee8e50' 2025-09-07T07:29:15.2547750Z Submodule path 'third_party/pocketfft': checked out '0fa0ef591e38c2758e3184c6c23e497b9f732ffa' 2025-09-07T07:29:15.4857227Z Submodule path 'third_party/protobuf': checked out 'd1eca4e4b421cd2997495c4b4e65cea6be4e9b8a' 2025-09-07T07:29:15.4879173Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:29:15.4881092Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/protobuf/third_party/googletest' 2025-09-07T07:29:15.4906403Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/benchmark'... 2025-09-07T07:29:16.9839208Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/googletest'... 2025-09-07T07:29:16.9963531Z Submodule path 'third_party/protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' 2025-09-07T07:29:17.0606511Z Submodule path 'third_party/protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' 2025-09-07T07:29:17.0700728Z Submodule path 'third_party/psimd': checked out '072586a71b55b7f8c584153d223e95687148a900' 2025-09-07T07:29:17.0819688Z Submodule path 'third_party/pthreadpool': checked out '4fe0e1e183925bf8cfa6aae24237e724a96479b8' 2025-09-07T07:29:17.1181499Z Submodule path 'third_party/pybind11': checked out 'f5fbe867d2d26e4a0a9177a51f6e568868ad3dc8' 2025-09-07T07:29:17.1438868Z Submodule path 'third_party/python-peachpy': checked out 'f45429b087dd7d5bc78bb40dc7cf06425c252d67' 2025-09-07T07:29:17.1834445Z Submodule path 'third_party/sleef': checked out '5a1d179df9cf652951b59010a2d2075372d67f68' 2025-09-07T07:29:17.2076578Z Submodule path 'third_party/tensorpipe': checked out 'af0118d13e52f5a08841464a768e01a0bf3e3075' 2025-09-07T07:29:17.2093624Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:29:17.2095746Z Submodule 'third_party/libnop' (https://github.com/google/libnop.git) registered for path 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:29:17.2097888Z Submodule 'third_party/libuv' (https://github.com/libuv/libuv.git) registered for path 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:29:17.2100116Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:29:17.2125457Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/googletest'... 2025-09-07T07:29:17.9241289Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libnop'... 2025-09-07T07:29:17.9616475Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libuv'... 2025-09-07T07:29:18.1770421Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11'... 2025-09-07T07:29:18.2271344Z Submodule path 'third_party/tensorpipe/third_party/googletest': checked out 'aee0f9d9b5b87796ee8a0ab26b7587ec30e8858e' 2025-09-07T07:29:18.2414241Z Submodule path 'third_party/tensorpipe/third_party/libnop': checked out '910b55815be16109f04f4180e9adee14fb4ce281' 2025-09-07T07:29:18.3062717Z Submodule path 'third_party/tensorpipe/third_party/libuv': checked out '5152db2cbfeb5582e9c27c5ea1dba2cd9e10759b' 2025-09-07T07:29:18.3328290Z Submodule path 'third_party/tensorpipe/third_party/pybind11': checked out 'a23996fce38ff6ccfbcdc09f1e63f2c4be5ea2ef' 2025-09-07T07:29:18.3342800Z Submodule 'tools/clang' (https://github.com/wjakob/clang-cindex-python3) registered for path 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:29:18.3367289Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11/tools/clang'... 2025-09-07T07:29:18.5139300Z Submodule path 'third_party/tensorpipe/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2025-09-07T07:29:18.5176532Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-09-07T07:29:18.5451426Z Entering 'android/libs/fbjni' 2025-09-07T07:29:18.5493063Z Entering 'third_party/FP16' 2025-09-07T07:29:18.5536179Z Entering 'third_party/FXdiv' 2025-09-07T07:29:18.5577868Z Entering 'third_party/NNPACK' 2025-09-07T07:29:18.5618606Z Entering 'third_party/NVTX' 2025-09-07T07:29:18.5660146Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:29:18.5700682Z Entering 'third_party/XNNPACK' 2025-09-07T07:29:18.5751669Z Entering 'third_party/aiter' 2025-09-07T07:29:18.5793062Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:29:18.5839593Z Entering 'third_party/benchmark' 2025-09-07T07:29:18.5881140Z Entering 'third_party/composable_kernel' 2025-09-07T07:29:18.5927272Z Entering 'third_party/cpp-httplib' 2025-09-07T07:29:18.5967953Z Entering 'third_party/cpuinfo' 2025-09-07T07:29:18.6009894Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:29:18.6051105Z Entering 'third_party/cutlass' 2025-09-07T07:29:18.6098365Z Entering 'third_party/fbgemm' 2025-09-07T07:29:18.6141270Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:29:18.6179621Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:29:18.6223484Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:29:18.6262645Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:29:18.6307646Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:29:18.6345444Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:29:18.6385208Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:29:18.6427528Z Entering 'third_party/flash-attention' 2025-09-07T07:29:18.6469000Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:29:18.6512480Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:29:18.6559594Z Entering 'third_party/flatbuffers' 2025-09-07T07:29:18.6603425Z Entering 'third_party/fmt' 2025-09-07T07:29:18.6644399Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:29:18.6685552Z Entering 'third_party/gloo' 2025-09-07T07:29:18.6726456Z Entering 'third_party/googletest' 2025-09-07T07:29:18.6767081Z Entering 'third_party/ideep' 2025-09-07T07:29:18.6805873Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:29:18.6851247Z Entering 'third_party/ittapi' 2025-09-07T07:29:18.6891980Z Entering 'third_party/kineto' 2025-09-07T07:29:18.6932053Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:29:18.6972366Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:29:18.7011680Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:29:18.7051118Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:29:18.7090036Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:29:18.7127612Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:29:18.7168939Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:29:18.7208123Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:29:18.7246732Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:29:18.7286406Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:29:18.7327670Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:29:18.7366060Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:29:18.7407102Z Entering 'third_party/kleidiai' 2025-09-07T07:29:18.7448387Z Entering 'third_party/mimalloc' 2025-09-07T07:29:18.7489102Z Entering 'third_party/nlohmann' 2025-09-07T07:29:18.7530121Z Entering 'third_party/onnx' 2025-09-07T07:29:18.7583253Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:29:18.7626890Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:29:18.7669590Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:29:18.7708715Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:29:18.7747828Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:29:18.7786460Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:29:18.7825097Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:29:18.7863890Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:29:18.7902727Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:29:18.7940953Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:29:18.7982291Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:29:18.8022889Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:29:18.8076846Z Entering 'third_party/pocketfft' 2025-09-07T07:29:18.8116757Z Entering 'third_party/protobuf' 2025-09-07T07:29:18.8159701Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:29:18.8200147Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:29:18.8240762Z Entering 'third_party/psimd' 2025-09-07T07:29:18.8280344Z Entering 'third_party/pthreadpool' 2025-09-07T07:29:18.8320862Z Entering 'third_party/pybind11' 2025-09-07T07:29:18.8360846Z Entering 'third_party/python-peachpy' 2025-09-07T07:29:18.8401173Z Entering 'third_party/sleef' 2025-09-07T07:29:18.8442249Z Entering 'third_party/tensorpipe' 2025-09-07T07:29:18.8482329Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:29:18.8520990Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:29:18.8559985Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:29:18.8599059Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:29:18.8636808Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:29:18.8690572Z ##[endgroup] 2025-09-07T07:29:18.8691016Z ##[group]Persisting credentials for submodules 2025-09-07T07:29:18.8697505Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-09-07T07:29:18.8968717Z Entering 'android/libs/fbjni' 2025-09-07T07:29:18.9021646Z Entering 'third_party/FP16' 2025-09-07T07:29:18.9074902Z Entering 'third_party/FXdiv' 2025-09-07T07:29:18.9127560Z Entering 'third_party/NNPACK' 2025-09-07T07:29:18.9180537Z Entering 'third_party/NVTX' 2025-09-07T07:29:18.9233703Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:29:18.9285438Z Entering 'third_party/XNNPACK' 2025-09-07T07:29:18.9346687Z Entering 'third_party/aiter' 2025-09-07T07:29:18.9399648Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:29:18.9456720Z Entering 'third_party/benchmark' 2025-09-07T07:29:18.9508889Z Entering 'third_party/composable_kernel' 2025-09-07T07:29:18.9565751Z Entering 'third_party/cpp-httplib' 2025-09-07T07:29:18.9617406Z Entering 'third_party/cpuinfo' 2025-09-07T07:29:18.9670292Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:29:18.9722842Z Entering 'third_party/cutlass' 2025-09-07T07:29:18.9781921Z Entering 'third_party/fbgemm' 2025-09-07T07:29:18.9834987Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:29:18.9886728Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:29:18.9941818Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:29:18.9992648Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:29:19.0049286Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:29:19.0099851Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:29:19.0149841Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:29:19.0203535Z Entering 'third_party/flash-attention' 2025-09-07T07:29:19.0255535Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:29:19.0310363Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:29:19.0368507Z Entering 'third_party/flatbuffers' 2025-09-07T07:29:19.0422111Z Entering 'third_party/fmt' 2025-09-07T07:29:19.0475260Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:29:19.0527388Z Entering 'third_party/gloo' 2025-09-07T07:29:19.0578997Z Entering 'third_party/googletest' 2025-09-07T07:29:19.0631964Z Entering 'third_party/ideep' 2025-09-07T07:29:19.0682423Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:29:19.0738530Z Entering 'third_party/ittapi' 2025-09-07T07:29:19.0790662Z Entering 'third_party/kineto' 2025-09-07T07:29:19.0841684Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:29:19.0891496Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:29:19.0943459Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:29:19.0994231Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:29:19.1046257Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:29:19.1097188Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:29:19.1149399Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:29:19.1199424Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:29:19.1249813Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:29:19.1300886Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:29:19.1355006Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:29:19.1406248Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:29:19.1458670Z Entering 'third_party/kleidiai' 2025-09-07T07:29:19.1510782Z Entering 'third_party/mimalloc' 2025-09-07T07:29:19.1562565Z Entering 'third_party/nlohmann' 2025-09-07T07:29:19.1616093Z Entering 'third_party/onnx' 2025-09-07T07:29:19.1679582Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:29:19.1734921Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:29:19.1788022Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:29:19.1838092Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:29:19.1887857Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:29:19.1937812Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:29:19.1989222Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:29:19.2039856Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:29:19.2090130Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:29:19.2139564Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:29:19.2192566Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:29:19.2244112Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:29:19.2308564Z Entering 'third_party/pocketfft' 2025-09-07T07:29:19.2360301Z Entering 'third_party/protobuf' 2025-09-07T07:29:19.2414034Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:29:19.2464753Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:29:19.2516926Z Entering 'third_party/psimd' 2025-09-07T07:29:19.2568892Z Entering 'third_party/pthreadpool' 2025-09-07T07:29:19.2620764Z Entering 'third_party/pybind11' 2025-09-07T07:29:19.2672955Z Entering 'third_party/python-peachpy' 2025-09-07T07:29:19.2725137Z Entering 'third_party/sleef' 2025-09-07T07:29:19.2776543Z Entering 'third_party/tensorpipe' 2025-09-07T07:29:19.2828517Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:29:19.2878558Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:29:19.2928797Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:29:19.2979054Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:29:19.3028738Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:29:19.3103194Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-09-07T07:29:19.3373204Z Entering 'android/libs/fbjni' 2025-09-07T07:29:19.3420673Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-09-07T07:29:19.3436950Z Entering 'third_party/FP16' 2025-09-07T07:29:19.3488079Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-09-07T07:29:19.3504102Z Entering 'third_party/FXdiv' 2025-09-07T07:29:19.3552115Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-09-07T07:29:19.3568065Z Entering 'third_party/NNPACK' 2025-09-07T07:29:19.3615750Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-09-07T07:29:19.3631824Z Entering 'third_party/NVTX' 2025-09-07T07:29:19.3680571Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-09-07T07:29:19.3697128Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:29:19.3745616Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-09-07T07:29:19.3761451Z Entering 'third_party/XNNPACK' 2025-09-07T07:29:19.3808913Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-09-07T07:29:19.3834051Z Entering 'third_party/aiter' 2025-09-07T07:29:19.3882322Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-09-07T07:29:19.3898479Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:29:19.3945425Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-09-07T07:29:19.3967644Z Entering 'third_party/benchmark' 2025-09-07T07:29:19.4015724Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-09-07T07:29:19.4031488Z Entering 'third_party/composable_kernel' 2025-09-07T07:29:19.4079145Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-09-07T07:29:19.4100679Z Entering 'third_party/cpp-httplib' 2025-09-07T07:29:19.4149203Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-09-07T07:29:19.4165151Z Entering 'third_party/cpuinfo' 2025-09-07T07:29:19.4213303Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-09-07T07:29:19.4229228Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:29:19.4278227Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-09-07T07:29:19.4294136Z Entering 'third_party/cutlass' 2025-09-07T07:29:19.4342928Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-09-07T07:29:19.4364749Z Entering 'third_party/fbgemm' 2025-09-07T07:29:19.4414181Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-09-07T07:29:19.4431208Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:29:19.4478183Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-09-07T07:29:19.4493729Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:29:19.4542224Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-09-07T07:29:19.4562148Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:29:19.4610460Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-09-07T07:29:19.4625898Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:29:19.4674752Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-09-07T07:29:19.4695880Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:29:19.4743216Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-09-07T07:29:19.4758023Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:29:19.4805959Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-09-07T07:29:19.4820672Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:29:19.4868944Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-09-07T07:29:19.4887114Z Entering 'third_party/flash-attention' 2025-09-07T07:29:19.4936374Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-09-07T07:29:19.4951787Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:29:19.4999235Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-09-07T07:29:19.5018614Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:29:19.5066514Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-09-07T07:29:19.5088963Z Entering 'third_party/flatbuffers' 2025-09-07T07:29:19.5137496Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-09-07T07:29:19.5155195Z Entering 'third_party/fmt' 2025-09-07T07:29:19.5203183Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-09-07T07:29:19.5218924Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:29:19.5268521Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-09-07T07:29:19.5284085Z Entering 'third_party/gloo' 2025-09-07T07:29:19.5332374Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-09-07T07:29:19.5348539Z Entering 'third_party/googletest' 2025-09-07T07:29:19.5397406Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-09-07T07:29:19.5413189Z Entering 'third_party/ideep' 2025-09-07T07:29:19.5462147Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-09-07T07:29:19.5476796Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:29:19.5523435Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-09-07T07:29:19.5544581Z Entering 'third_party/ittapi' 2025-09-07T07:29:19.5593311Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-09-07T07:29:19.5608976Z Entering 'third_party/kineto' 2025-09-07T07:29:19.5657146Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-09-07T07:29:19.5672589Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:29:19.5721651Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-09-07T07:29:19.5735943Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:29:19.5784161Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-09-07T07:29:19.5800286Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:29:19.5849274Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-09-07T07:29:19.5864659Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:29:19.5913429Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-09-07T07:29:19.5928669Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:29:19.5976460Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-09-07T07:29:19.5990180Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:29:19.6039237Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-09-07T07:29:19.6056314Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:29:19.6104516Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-09-07T07:29:19.6119626Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:29:19.6168414Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-09-07T07:29:19.6183684Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:29:19.6232544Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-09-07T07:29:19.6248301Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:29:19.6297474Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-09-07T07:29:19.6314393Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:29:19.6362986Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-09-07T07:29:19.6378346Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:29:19.6426465Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-09-07T07:29:19.6443503Z Entering 'third_party/kleidiai' 2025-09-07T07:29:19.6492349Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-09-07T07:29:19.6508462Z Entering 'third_party/mimalloc' 2025-09-07T07:29:19.6556590Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-09-07T07:29:19.6572518Z Entering 'third_party/nlohmann' 2025-09-07T07:29:19.6621088Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-09-07T07:29:19.6638058Z Entering 'third_party/onnx' 2025-09-07T07:29:19.6687009Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-09-07T07:29:19.6713986Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:29:19.6762048Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-09-07T07:29:19.6781034Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:29:19.6830513Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-09-07T07:29:19.6846669Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:29:19.6892911Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-09-07T07:29:19.6907989Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:29:19.6956370Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-09-07T07:29:19.6971681Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:29:19.7019156Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-09-07T07:29:19.7034044Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:29:19.7081789Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-09-07T07:29:19.7097494Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:29:19.7146521Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-09-07T07:29:19.7161454Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:29:19.7210229Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-09-07T07:29:19.7225190Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:29:19.7273979Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-09-07T07:29:19.7287972Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:29:19.7335801Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-09-07T07:29:19.7352134Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:29:19.7400112Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-09-07T07:29:19.7417230Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:29:19.7464586Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-09-07T07:29:19.7492995Z Entering 'third_party/pocketfft' 2025-09-07T07:29:19.7541504Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-09-07T07:29:19.7556883Z Entering 'third_party/protobuf' 2025-09-07T07:29:19.7605438Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-09-07T07:29:19.7622835Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:29:19.7671799Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-09-07T07:29:19.7687153Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:29:19.7734858Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-09-07T07:29:19.7752103Z Entering 'third_party/psimd' 2025-09-07T07:29:19.7799837Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-09-07T07:29:19.7815505Z Entering 'third_party/pthreadpool' 2025-09-07T07:29:19.7863869Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-09-07T07:29:19.7879645Z Entering 'third_party/pybind11' 2025-09-07T07:29:19.7927650Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-09-07T07:29:19.7943799Z Entering 'third_party/python-peachpy' 2025-09-07T07:29:19.7990855Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-09-07T07:29:19.8006618Z Entering 'third_party/sleef' 2025-09-07T07:29:19.8053805Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-09-07T07:29:19.8069624Z Entering 'third_party/tensorpipe' 2025-09-07T07:29:19.8119541Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-09-07T07:29:19.8134769Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:29:19.8182893Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-09-07T07:29:19.8198496Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:29:19.8245561Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-09-07T07:29:19.8260333Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:29:19.8307549Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-09-07T07:29:19.8322959Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:29:19.8370900Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-09-07T07:29:19.8385194Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:29:19.8433991Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-09-07T07:29:19.8907285Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-09-07T07:29:19.9177394Z Entering 'android/libs/fbjni' 2025-09-07T07:29:19.9217484Z Entering 'third_party/FP16' 2025-09-07T07:29:19.9258360Z Entering 'third_party/FXdiv' 2025-09-07T07:29:19.9297937Z Entering 'third_party/NNPACK' 2025-09-07T07:29:19.9338959Z Entering 'third_party/NVTX' 2025-09-07T07:29:19.9379446Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:29:19.9419963Z Entering 'third_party/XNNPACK' 2025-09-07T07:29:19.9469263Z Entering 'third_party/aiter' 2025-09-07T07:29:19.9509713Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:29:19.9554166Z Entering 'third_party/benchmark' 2025-09-07T07:29:19.9595580Z Entering 'third_party/composable_kernel' 2025-09-07T07:29:19.9641793Z Entering 'third_party/cpp-httplib' 2025-09-07T07:29:19.9681792Z Entering 'third_party/cpuinfo' 2025-09-07T07:29:19.9722285Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:29:19.9762564Z Entering 'third_party/cutlass' 2025-09-07T07:29:19.9809037Z Entering 'third_party/fbgemm' 2025-09-07T07:29:19.9851158Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:29:19.9890615Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:29:19.9935589Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:29:19.9974763Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:29:20.0018906Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:29:20.0058110Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:29:20.0097551Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:29:20.0138762Z Entering 'third_party/flash-attention' 2025-09-07T07:29:20.0179816Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:29:20.0222736Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:29:20.0268230Z Entering 'third_party/flatbuffers' 2025-09-07T07:29:20.0310507Z Entering 'third_party/fmt' 2025-09-07T07:29:20.0350549Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:29:20.0390759Z Entering 'third_party/gloo' 2025-09-07T07:29:20.0430812Z Entering 'third_party/googletest' 2025-09-07T07:29:20.0471486Z Entering 'third_party/ideep' 2025-09-07T07:29:20.0511520Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:29:20.0556065Z Entering 'third_party/ittapi' 2025-09-07T07:29:20.0596895Z Entering 'third_party/kineto' 2025-09-07T07:29:20.0637888Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:29:20.0676666Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:29:20.0717032Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:29:20.0756282Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:29:20.0796033Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:29:20.0834354Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:29:20.0876077Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:29:20.0914293Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:29:20.0954351Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:29:20.0994343Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:29:20.1035898Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:29:20.1075100Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:29:20.1115397Z Entering 'third_party/kleidiai' 2025-09-07T07:29:20.1155497Z Entering 'third_party/mimalloc' 2025-09-07T07:29:20.1196448Z Entering 'third_party/nlohmann' 2025-09-07T07:29:20.1238055Z Entering 'third_party/onnx' 2025-09-07T07:29:20.1290122Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:29:20.1332157Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:29:20.1372944Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:29:20.1411981Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:29:20.1450561Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:29:20.1488970Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:29:20.1528262Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:29:20.1567031Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:29:20.1605400Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:29:20.1643539Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:29:20.1684931Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:29:20.1726061Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:29:20.1780012Z Entering 'third_party/pocketfft' 2025-09-07T07:29:20.1819859Z Entering 'third_party/protobuf' 2025-09-07T07:29:20.1861663Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:29:20.1900733Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:29:20.1943030Z Entering 'third_party/psimd' 2025-09-07T07:29:20.1982676Z Entering 'third_party/pthreadpool' 2025-09-07T07:29:20.2022689Z Entering 'third_party/pybind11' 2025-09-07T07:29:20.2062983Z Entering 'third_party/python-peachpy' 2025-09-07T07:29:20.2103260Z Entering 'third_party/sleef' 2025-09-07T07:29:20.2143662Z Entering 'third_party/tensorpipe' 2025-09-07T07:29:20.2184069Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:29:20.2222557Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:29:20.2260922Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:29:20.2299188Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:29:20.2336855Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:29:20.2395126Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-09-07T07:29:20.2666149Z Entering 'android/libs/fbjni' 2025-09-07T07:29:20.2706773Z Entering 'third_party/FP16' 2025-09-07T07:29:20.2747443Z Entering 'third_party/FXdiv' 2025-09-07T07:29:20.2787908Z Entering 'third_party/NNPACK' 2025-09-07T07:29:20.2829205Z Entering 'third_party/NVTX' 2025-09-07T07:29:20.2872148Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:29:20.2912198Z Entering 'third_party/XNNPACK' 2025-09-07T07:29:20.2962548Z Entering 'third_party/aiter' 2025-09-07T07:29:20.3004001Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:29:20.3049722Z Entering 'third_party/benchmark' 2025-09-07T07:29:20.3091182Z Entering 'third_party/composable_kernel' 2025-09-07T07:29:20.3137052Z Entering 'third_party/cpp-httplib' 2025-09-07T07:29:20.3177497Z Entering 'third_party/cpuinfo' 2025-09-07T07:29:20.3218464Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:29:20.3259823Z Entering 'third_party/cutlass' 2025-09-07T07:29:20.3306679Z Entering 'third_party/fbgemm' 2025-09-07T07:29:20.3348894Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:29:20.3387695Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:29:20.3431858Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:29:20.3470598Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:29:20.3515848Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:29:20.3554261Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:29:20.3592668Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:29:20.3634661Z Entering 'third_party/flash-attention' 2025-09-07T07:29:20.3675439Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:29:20.3718418Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:29:20.3764391Z Entering 'third_party/flatbuffers' 2025-09-07T07:29:20.3807550Z Entering 'third_party/fmt' 2025-09-07T07:29:20.3848178Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:29:20.3889044Z Entering 'third_party/gloo' 2025-09-07T07:29:20.3929162Z Entering 'third_party/googletest' 2025-09-07T07:29:20.3969693Z Entering 'third_party/ideep' 2025-09-07T07:29:20.4009279Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:29:20.4054930Z Entering 'third_party/ittapi' 2025-09-07T07:29:20.4095194Z Entering 'third_party/kineto' 2025-09-07T07:29:20.4134445Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:29:20.4173479Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:29:20.4214040Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:29:20.4253027Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:29:20.4293252Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:29:20.4332469Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:29:20.4374073Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:29:20.4413922Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:29:20.4453242Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:29:20.4492667Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:29:20.4533980Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:29:20.4572386Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:29:20.4612937Z Entering 'third_party/kleidiai' 2025-09-07T07:29:20.4653939Z Entering 'third_party/mimalloc' 2025-09-07T07:29:20.4695506Z Entering 'third_party/nlohmann' 2025-09-07T07:29:20.4736918Z Entering 'third_party/onnx' 2025-09-07T07:29:20.4788824Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:29:20.4830911Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:29:20.4872768Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:29:20.4910851Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:29:20.4949929Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:29:20.4989821Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:29:20.5029834Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:29:20.5068381Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:29:20.5106179Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:29:20.5143719Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:29:20.5184769Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:29:20.5225356Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:29:20.5278355Z Entering 'third_party/pocketfft' 2025-09-07T07:29:20.5319063Z Entering 'third_party/protobuf' 2025-09-07T07:29:20.5360665Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:29:20.5399828Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:29:20.5441921Z Entering 'third_party/psimd' 2025-09-07T07:29:20.5482172Z Entering 'third_party/pthreadpool' 2025-09-07T07:29:20.5521947Z Entering 'third_party/pybind11' 2025-09-07T07:29:20.5562340Z Entering 'third_party/python-peachpy' 2025-09-07T07:29:20.5603044Z Entering 'third_party/sleef' 2025-09-07T07:29:20.5643056Z Entering 'third_party/tensorpipe' 2025-09-07T07:29:20.5683260Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:29:20.5721012Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:29:20.5759136Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:29:20.5797746Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:29:20.5836131Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:29:20.5890603Z ##[endgroup] 2025-09-07T07:29:20.5925040Z [command]/usr/bin/git log -1 --format=%H 2025-09-07T07:29:20.5947251Z 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:29:20.6050448Z ##[group]Run cd "${GITHUB_WORKSPACE}" 2025-09-07T07:29:20.6050764Z cd "${GITHUB_WORKSPACE}" 2025-09-07T07:29:20.6051032Z # Clean stale submodule dirs 2025-09-07T07:29:20.6051299Z if [ -z "${NO_SUDO}" ]; then 2025-09-07T07:29:20.6051629Z  sudo git submodule foreach --recursive git clean -ffdx 2025-09-07T07:29:20.6051957Z else 2025-09-07T07:29:20.6052210Z  git submodule foreach --recursive git clean -ffdx 2025-09-07T07:29:20.6052528Z fi 2025-09-07T07:29:20.6062857Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:29:20.6063173Z env: 2025-09-07T07:29:20.6063349Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:20.6063582Z NO_SUDO: true 2025-09-07T07:29:20.6063758Z ##[endgroup] 2025-09-07T07:29:20.6393130Z Entering 'android/libs/fbjni' 2025-09-07T07:29:20.6424866Z Entering 'third_party/FP16' 2025-09-07T07:29:20.6455012Z Entering 'third_party/FXdiv' 2025-09-07T07:29:20.6485000Z Entering 'third_party/NNPACK' 2025-09-07T07:29:20.6518500Z Entering 'third_party/NVTX' 2025-09-07T07:29:20.6554699Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:29:20.6586186Z Entering 'third_party/XNNPACK' 2025-09-07T07:29:20.6691109Z Entering 'third_party/aiter' 2025-09-07T07:29:20.6729589Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:29:20.6822608Z Entering 'third_party/benchmark' 2025-09-07T07:29:20.6853784Z Entering 'third_party/composable_kernel' 2025-09-07T07:29:20.6954781Z Entering 'third_party/cpp-httplib' 2025-09-07T07:29:20.6986807Z Entering 'third_party/cpuinfo' 2025-09-07T07:29:20.7021538Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:29:20.7054503Z Entering 'third_party/cutlass' 2025-09-07T07:29:20.7140272Z Entering 'third_party/fbgemm' 2025-09-07T07:29:20.7194087Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:29:20.7224011Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:29:20.7315041Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:29:20.7346787Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:29:20.7429963Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:29:20.7461049Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:29:20.7488560Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:29:20.7529379Z Entering 'third_party/flash-attention' 2025-09-07T07:29:20.7566826Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:29:20.7650970Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:29:20.7727219Z Entering 'third_party/flatbuffers' 2025-09-07T07:29:20.7789226Z Entering 'third_party/fmt' 2025-09-07T07:29:20.7820215Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:29:20.7852082Z Entering 'third_party/gloo' 2025-09-07T07:29:20.7884381Z Entering 'third_party/googletest' 2025-09-07T07:29:20.7916757Z Entering 'third_party/ideep' 2025-09-07T07:29:20.7945014Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:29:20.8018076Z Entering 'third_party/ittapi' 2025-09-07T07:29:20.8049991Z Entering 'third_party/kineto' 2025-09-07T07:29:20.8082673Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:29:20.8114224Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:29:20.8155119Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:29:20.8185049Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:29:20.8215837Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:29:20.8243403Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:29:20.8272763Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:29:20.8302038Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:29:20.8333202Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:29:20.8370476Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:29:20.8401231Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:29:20.8430772Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:29:20.8462750Z Entering 'third_party/kleidiai' 2025-09-07T07:29:20.8497451Z Entering 'third_party/mimalloc' 2025-09-07T07:29:20.8529299Z Entering 'third_party/nlohmann' 2025-09-07T07:29:20.8571039Z Entering 'third_party/onnx' 2025-09-07T07:29:20.8837060Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:29:20.8871716Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:29:20.8921128Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:29:20.8950681Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:29:20.8981308Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:29:20.9009300Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:29:20.9047051Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:29:20.9076916Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:29:20.9106139Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:29:20.9135124Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:29:20.9177262Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:29:20.9209751Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:29:20.9424012Z Entering 'third_party/pocketfft' 2025-09-07T07:29:20.9453256Z Entering 'third_party/protobuf' 2025-09-07T07:29:20.9521002Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:29:20.9549961Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:29:20.9583817Z Entering 'third_party/psimd' 2025-09-07T07:29:20.9613138Z Entering 'third_party/pthreadpool' 2025-09-07T07:29:20.9642605Z Entering 'third_party/pybind11' 2025-09-07T07:29:20.9675280Z Entering 'third_party/python-peachpy' 2025-09-07T07:29:20.9706327Z Entering 'third_party/sleef' 2025-09-07T07:29:20.9738282Z Entering 'third_party/tensorpipe' 2025-09-07T07:29:20.9770473Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:29:20.9800518Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:29:20.9829669Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:29:20.9862383Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:29:20.9890594Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:29:21.0041173Z Prepare all required actions 2025-09-07T07:29:21.0041731Z Getting action download info 2025-09-07T07:29:21.1315685Z ##[group]Run ./.github/actions/setup-linux 2025-09-07T07:29:21.1315982Z env: 2025-09-07T07:29:21.1316159Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:21.1316366Z ##[endgroup] 2025-09-07T07:29:21.1357211Z ##[group]Run set -euo pipefail 2025-09-07T07:29:21.1357504Z set -euo pipefail 2025-09-07T07:29:21.1357749Z function get_ec2_metadata() { 2025-09-07T07:29:21.1358071Z  # Pulled from instance metadata endpoint for EC2 2025-09-07T07:29:21.1358617Z  # see https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/instancedata-data-retrieval.html 2025-09-07T07:29:21.1359100Z  category=$1 2025-09-07T07:29:21.1359405Z  # If it is GCP runner (runner name contains gcp), do not run this 2025-09-07T07:29:21.1360017Z  runner_name_str=i-0c7ee175fda0aa6a4 2025-09-07T07:29:21.1360332Z  if [[ -f /.inarc ]]; then 2025-09-07T07:29:21.1360631Z  echo "ARC Runner, no info on ec2 metadata" 2025-09-07T07:29:21.1360961Z  elif [[ $runner_name_str == *"gcp"* ]]; then 2025-09-07T07:29:21.1361365Z  echo "Runner is from Google Cloud Platform, No info on ec2 metadata" 2025-09-07T07:29:21.1361728Z  else 2025-09-07T07:29:21.1362474Z  curl -H "X-aws-ec2-metadata-token: $(curl -s -X PUT "http://169.254.169.254/latest/api/token" -H "X-aws-ec2-metadata-token-ttl-seconds: 30")" -fsSL "http://169.254.169.254/latest/meta-data/${category}" 2025-09-07T07:29:21.1363258Z  fi 2025-09-07T07:29:21.1363431Z } 2025-09-07T07:29:21.1363652Z echo "ami-id: $(get_ec2_metadata ami-id)" 2025-09-07T07:29:21.1364015Z echo "instance-id: $(get_ec2_metadata instance-id)" 2025-09-07T07:29:21.1364433Z echo "instance-type: $(get_ec2_metadata instance-type)" 2025-09-07T07:29:21.1364785Z echo "system info $(uname -a)" 2025-09-07T07:29:21.1372666Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:29:21.1372982Z env: 2025-09-07T07:29:21.1373158Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:21.1373369Z ##[endgroup] 2025-09-07T07:29:21.1502646Z ami-id: ami-022bbd2ccaf21691f 2025-09-07T07:29:21.1591050Z instance-id: i-0c7ee175fda0aa6a4 2025-09-07T07:29:21.1678859Z instance-type: m7g.metal 2025-09-07T07:29:21.1689098Z system info Linux ip-10-0-72-124.ec2.internal 6.1.141-155.222.amzn2023.aarch64 #1 SMP Tue Jun 17 10:29:19 UTC 2025 aarch64 aarch64 aarch64 GNU/Linux 2025-09-07T07:29:21.1708122Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:29:21.1708964Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:29:21.1715361Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:29:21.1715717Z env: 2025-09-07T07:29:21.1715915Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:21.1716156Z ##[endgroup] 2025-09-07T07:29:21.1762720Z ##[group]Run if systemctl is-active --quiet docker; then 2025-09-07T07:29:21.1763103Z if systemctl is-active --quiet docker; then 2025-09-07T07:29:21.1763437Z  echo "Docker daemon is running..."; 2025-09-07T07:29:21.1763706Z else 2025-09-07T07:29:21.1764001Z  echo "Starting docker daemon..." && sudo systemctl start docker; 2025-09-07T07:29:21.1764354Z fi 2025-09-07T07:29:21.1769886Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:29:21.1770202Z env: 2025-09-07T07:29:21.1770377Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:21.1770592Z ##[endgroup] 2025-09-07T07:29:21.1840665Z Docker daemon is running... 2025-09-07T07:29:21.1881546Z ##[group]Run nick-fields/retry@v3.0.0 2025-09-07T07:29:21.1881850Z with: 2025-09-07T07:29:21.1882064Z shell: bash 2025-09-07T07:29:21.1882616Z timeout_minutes: 5 2025-09-07T07:29:21.1882890Z max_attempts: 3 2025-09-07T07:29:21.1883103Z retry_wait_seconds: 30 2025-09-07T07:29:21.1885234Z command: AWS_ACCOUNT_ID=$(aws sts get-caller-identity|grep Account|cut -f4 -d\") aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" # For LF Runners we need to make sure we also login to Meta's ECR docker registry too. META_AWS_ACCOUNT_ID=308535385114 if [ "$AWS_ACCOUNT_ID" != "$META_AWS_ACCOUNT_ID" ] ; then aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$META_AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" fi 2025-09-07T07:29:21.1887301Z polling_interval_seconds: 1 2025-09-07T07:29:21.1887787Z warning_on_retry: true 2025-09-07T07:29:21.1888070Z continue_on_error: false 2025-09-07T07:29:21.1888319Z env: 2025-09-07T07:29:21.1888554Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:21.1888840Z AWS_RETRY_MODE: standard 2025-09-07T07:29:21.1889104Z AWS_MAX_ATTEMPTS: 5 2025-09-07T07:29:21.1889408Z AWS_DEFAULT_REGION: us-east-1 2025-09-07T07:29:21.1889670Z ##[endgroup] 2025-09-07T07:29:22.1980085Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-09-07T07:29:22.1980829Z Configure a credential helper to remove this warning. See 2025-09-07T07:29:22.1981385Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-09-07T07:29:22.1981754Z 2025-09-07T07:29:22.1981850Z Login Succeeded 2025-09-07T07:29:22.2633419Z Command completed after 1 attempt(s). 2025-09-07T07:29:22.2685190Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:29:22.2685710Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:29:22.2686193Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:29:22.2693875Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:29:22.2694254Z env: 2025-09-07T07:29:22.2694593Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:22.2694850Z ##[endgroup] 2025-09-07T07:29:22.2778674Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-09-07T07:29:22.2779157Z # ignore expansion of "docker ps -q" since it could be empty 2025-09-07T07:29:22.2779520Z # shellcheck disable=SC2046 2025-09-07T07:29:22.2779801Z docker stop $(docker ps -q) || true 2025-09-07T07:29:22.2780096Z # Prune all of the docker images 2025-09-07T07:29:22.2780366Z docker system prune -af 2025-09-07T07:29:22.2786295Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:29:22.2786608Z env: 2025-09-07T07:29:22.2786782Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:22.2787014Z ##[endgroup] 2025-09-07T07:29:22.3193629Z "docker stop" requires at least 1 argument. 2025-09-07T07:29:22.3193930Z See 'docker stop --help'. 2025-09-07T07:29:22.3194078Z 2025-09-07T07:29:22.3194236Z Usage: docker stop [OPTIONS] CONTAINER [CONTAINER...] 2025-09-07T07:29:22.3194476Z 2025-09-07T07:29:22.3194568Z Stop one or more running containers 2025-09-07T07:29:22.3347036Z Total reclaimed space: 0B 2025-09-07T07:29:22.3376760Z ##[group]Run set +e 2025-09-07T07:29:22.3377002Z set +e 2025-09-07T07:29:22.3377187Z set -x 2025-09-07T07:29:22.3377366Z  2025-09-07T07:29:22.3377565Z PT_DOMAIN=download.pytorch.org 2025-09-07T07:29:22.3378065Z # TODO: Flaky access to download.pytorch.org https://github.com/pytorch/pytorch/issues/100400, 2025-09-07T07:29:22.3378837Z # cleaning this up once the issue is fixed. There are more than one resolved IP here, the last 2025-09-07T07:29:22.3379332Z # one is returned at random 2025-09-07T07:29:22.3379685Z RESOLVED_IP=$(dig -4 +short "${PT_DOMAIN}" | tail -n1) 2025-09-07T07:29:22.3379996Z  2025-09-07T07:29:22.3380186Z if [ -z "${RESOLVED_IP}" ]; then 2025-09-07T07:29:22.3380881Z  echo "Couldn't resolve ${PT_DOMAIN}, retrying with Google DNS..." 2025-09-07T07:29:22.3381361Z  RESOLVED_IP=$(dig -4 +short "${PT_DOMAIN}" @8.8.8.8 | tail -n1) 2025-09-07T07:29:22.3381692Z  2025-09-07T07:29:22.3381929Z  if [ -z "${RESOLVED_IP}" ]; then 2025-09-07T07:29:22.3382295Z  echo "Couldn't resolve ${PT_DOMAIN}, exiting..." 2025-09-07T07:29:22.3382602Z  exit 1 2025-09-07T07:29:22.3382791Z  fi 2025-09-07T07:29:22.3382957Z fi 2025-09-07T07:29:22.3383120Z  2025-09-07T07:29:22.3383333Z if grep -r "${PT_DOMAIN}" /etc/hosts; then 2025-09-07T07:29:22.3383645Z  # Clean up any old records first 2025-09-07T07:29:22.3383946Z  sudo sed -i "/${PT_DOMAIN}/d" /etc/hosts 2025-09-07T07:29:22.3384382Z fi 2025-09-07T07:29:22.3384548Z  2025-09-07T07:29:22.3384816Z echo "${RESOLVED_IP} ${PT_DOMAIN}" | sudo tee -a /etc/hosts 2025-09-07T07:29:22.3385147Z cat /etc/hosts 2025-09-07T07:29:22.3391417Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:29:22.3391739Z env: 2025-09-07T07:29:22.3391915Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:22.3392120Z ##[endgroup] 2025-09-07T07:29:22.3413799Z + PT_DOMAIN=download.pytorch.org 2025-09-07T07:29:22.3418600Z ++ dig -4 +short download.pytorch.org 2025-09-07T07:29:22.3419329Z ++ tail -n1 2025-09-07T07:29:22.3887253Z + RESOLVED_IP=18.160.10.22 2025-09-07T07:29:22.3887559Z + '[' -z 18.160.10.22 ']' 2025-09-07T07:29:22.3887811Z + grep -r download.pytorch.org /etc/hosts 2025-09-07T07:29:22.3901871Z + echo '18.160.10.22 download.pytorch.org' 2025-09-07T07:29:22.3902318Z + sudo tee -a /etc/hosts 2025-09-07T07:29:22.6617015Z 18.160.10.22 download.pytorch.org 2025-09-07T07:29:22.6633183Z + cat /etc/hosts 2025-09-07T07:29:22.6641475Z 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 2025-09-07T07:29:22.6646188Z ::1 localhost6 localhost6.localdomain6 2025-09-07T07:29:22.6646512Z 18.160.10.22 download.pytorch.org 2025-09-07T07:29:22.6795607Z ##[group]Run pytorch/test-infra/.github/actions/calculate-docker-image@main 2025-09-07T07:29:22.6796067Z with: 2025-09-07T07:29:22.6796848Z docker-image-name: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:22.6797717Z use-custom-docker-registry: true 2025-09-07T07:29:22.6797973Z docker-build-dir: .ci/docker 2025-09-07T07:29:22.6798218Z docker-build-script: ./build.sh 2025-09-07T07:29:22.6798465Z working-directory: . 2025-09-07T07:29:22.6798758Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:29:22.6799090Z force-push: false 2025-09-07T07:29:22.6799292Z env: 2025-09-07T07:29:22.6799462Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:22.6799671Z ##[endgroup] 2025-09-07T07:29:22.6816460Z ##[group]Run set -ex 2025-09-07T07:29:22.6816722Z set -ex 2025-09-07T07:29:22.6816906Z  2025-09-07T07:29:22.6817261Z # If the docker build directory or the build script doesn't exist, the action will 2025-09-07T07:29:22.6817843Z # gracefully return the docker image name as it is. Pulling docker image in Linux 2025-09-07T07:29:22.6818326Z # job could then download the pre-built image as usual 2025-09-07T07:29:22.6818925Z if [[ -d "${DOCKER_BUILD_DIR}" ]] && [[ -f "${DOCKER_BUILD_DIR}/${DOCKER_BUILD_SCRIPT}" ]] && [[ "${USE_CUSTOM_DOCKER_REGISTRY}" == "true" ]]; then 2025-09-07T07:29:22.6819480Z  echo "skip=false" >> "${GITHUB_OUTPUT}" 2025-09-07T07:29:22.6819757Z else 2025-09-07T07:29:22.6819978Z  echo "skip=true" >> "${GITHUB_OUTPUT}" 2025-09-07T07:29:22.6820363Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:29:22.6820706Z  2025-09-07T07:29:22.6821182Z  echo "Not using custom ECR registry. Either it was not requested or there is no Docker build script in the ${REPO_NAME} repo..." 2025-09-07T07:29:22.6821726Z  exit 0 2025-09-07T07:29:22.6821905Z fi 2025-09-07T07:29:22.6822073Z  2025-09-07T07:29:22.6822361Z if [[ "${DOCKER_IMAGE_NAME}" == *"${DOCKER_REGISTRY}/${REPO_NAME}"* ]]; then 2025-09-07T07:29:22.6822877Z  # The docker image name already includes the ECR prefix and tag, so we can just 2025-09-07T07:29:22.6823333Z  # use it as it is, but first let's extract the tag 2025-09-07T07:29:22.6823743Z  DOCKER_TAG=$(echo "${DOCKER_IMAGE_NAME}" | awk -F '[:,]' '{print $2}') 2025-09-07T07:29:22.6824177Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:29:22.6824598Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:29:22.6825166Z else 2025-09-07T07:29:22.6825389Z  if [[ "${DOCKER_IMAGE_NAME}" == *:* ]]; then 2025-09-07T07:29:22.6825715Z  CUSTOM_TAG_PREFIX=${DOCKER_IMAGE_NAME#*:} 2025-09-07T07:29:22.6826054Z  DOCKER_IMAGE_NAME=${DOCKER_IMAGE_NAME%%:*} 2025-09-07T07:29:22.6826336Z  fi 2025-09-07T07:29:22.6826721Z  DOCKER_TAG=${CUSTOM_TAG_PREFIX:+${CUSTOM_TAG_PREFIX}-}$(git rev-parse HEAD:"${DOCKER_BUILD_DIR}") 2025-09-07T07:29:22.6827243Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:29:22.6827795Z  echo "docker-image=${DOCKER_REGISTRY}/${REPO_NAME}/${DOCKER_IMAGE_NAME}:${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:29:22.6828400Z  echo "custom-tag-prefix=${CUSTOM_TAG_PREFIX}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:29:22.6828762Z fi 2025-09-07T07:29:22.6837576Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:29:22.6837911Z env: 2025-09-07T07:29:22.6838089Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:22.6838309Z REPO_NAME: pytorch 2025-09-07T07:29:22.6839265Z DOCKER_IMAGE_NAME: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:22.6840127Z DOCKER_BUILD_DIR: .ci/docker 2025-09-07T07:29:22.6840363Z DOCKER_BUILD_SCRIPT: ./build.sh 2025-09-07T07:29:22.6840684Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:29:22.6841023Z USE_CUSTOM_DOCKER_REGISTRY: true 2025-09-07T07:29:22.6841266Z CUSTOM_TAG_PREFIX: 2025-09-07T07:29:22.6841459Z ##[endgroup] 2025-09-07T07:29:22.6864457Z + [[ -d .ci/docker ]] 2025-09-07T07:29:22.6864684Z + [[ -f .ci/docker/./build.sh ]] 2025-09-07T07:29:22.6864925Z + [[ true == \t\r\u\e ]] 2025-09-07T07:29:22.6865134Z + echo skip=false 2025-09-07T07:29:22.6866205Z + [[ 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 == *\3\0\8\5\3\5\3\8\5\1\1\4\.\d\k\r\.\e\c\r\.\u\s\-\e\a\s\t\-\1\.\a\m\a\z\o\n\a\w\s\.\c\o\m\/\p\y\t\o\r\c\h* ]] 2025-09-07T07:29:22.6872501Z ++ echo 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:22.6873302Z ++ awk -F '[:,]' '{print $2}' 2025-09-07T07:29:22.6964876Z + DOCKER_TAG=pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:22.6966000Z + echo docker-tag=pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:22.6967340Z + echo docker-image=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:22.6987686Z ##[group]Run set +e 2025-09-07T07:29:22.6987930Z set +e 2025-09-07T07:29:22.6988117Z set -x 2025-09-07T07:29:22.6988297Z  2025-09-07T07:29:22.6988461Z login() { 2025-09-07T07:29:22.6988878Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-09-07T07:29:22.6989319Z } 2025-09-07T07:29:22.6989489Z  2025-09-07T07:29:22.6989653Z retry () { 2025-09-07T07:29:22.6989881Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-09-07T07:29:22.6990142Z } 2025-09-07T07:29:22.6990309Z  2025-09-07T07:29:22.6990500Z retry login "${DOCKER_REGISTRY}" 2025-09-07T07:29:22.6990756Z  2025-09-07T07:29:22.6990934Z START_TIME=$(date +%s) 2025-09-07T07:29:22.6991178Z # Wait up to 120 minutes 2025-09-07T07:29:22.6991484Z while [[ $(( $(date +%s) - 7200 )) -lt $START_TIME ]]; do 2025-09-07T07:29:22.6991898Z  # Check if image already exists, if it does then skip building it 2025-09-07T07:29:22.6992537Z  if docker manifest inspect "${DOCKER_IMAGE}"; then 2025-09-07T07:29:22.6992839Z  exit 0 2025-09-07T07:29:22.6993027Z  fi 2025-09-07T07:29:22.6993199Z  2025-09-07T07:29:22.6993528Z  # NB: This flag is used by Docker build workflow to push the image to ECR, so we can 2025-09-07T07:29:22.6994101Z  # use this to differentiate between the Docker build and regular build jobs. For the 2025-09-07T07:29:22.6994674Z  # latter, it will wait for the Docker images to become available before continuing 2025-09-07T07:29:22.6995120Z  if [ "${DOCKER_PUSH:-false}" == "true" ]; then 2025-09-07T07:29:22.6995461Z  # It's a Docker build job, let's build the image 2025-09-07T07:29:22.6995754Z  break 2025-09-07T07:29:22.6995942Z  else 2025-09-07T07:29:22.6996229Z  # It's a regular build job, wait for the image to become available 2025-09-07T07:29:22.6996586Z  sleep 300 2025-09-07T07:29:22.6996788Z  fi 2025-09-07T07:29:22.6996960Z done 2025-09-07T07:29:22.6997133Z  2025-09-07T07:29:22.6997592Z # NB: This part requires a full checkout. Otherwise, the merge base will 2025-09-07T07:29:22.6998100Z # be empty. The default action would be to continue rebuild the image 2025-09-07T07:29:22.6998543Z if [[ "$BASE_REVISION" = "$(git rev-parse HEAD)" ]]; then 2025-09-07T07:29:22.6998936Z  # if we're on the base branch then use the parent commit 2025-09-07T07:29:22.6999278Z  MERGE_BASE=$(git rev-parse HEAD~) 2025-09-07T07:29:22.6999536Z else 2025-09-07T07:29:22.6999810Z  # otherwise we're on a PR, so use the most recent base commit 2025-09-07T07:29:22.7000215Z  MERGE_BASE=$(git merge-base HEAD "$BASE_REVISION") 2025-09-07T07:29:22.7000515Z fi 2025-09-07T07:29:22.7000685Z  2025-09-07T07:29:22.7000875Z if [[ -z "${MERGE_BASE}" ]]; then 2025-09-07T07:29:22.7001182Z  echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-09-07T07:29:22.7001459Z  2025-09-07T07:29:22.7001863Z  echo "Finding merge base only works with full checkout, please set fetch-depth to 0, continuing ..." 2025-09-07T07:29:22.7002337Z  exit 0 2025-09-07T07:29:22.7002517Z fi 2025-09-07T07:29:22.7002686Z  2025-09-07T07:29:22.7002945Z if ! git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}"; then 2025-09-07T07:29:22.7003530Z  echo "Directory '${DOCKER_BUILD_DIR}' not found in commit $MERGE_BASE, you should rebase onto a more recent commit" 2025-09-07T07:29:22.7004025Z  exit 1 2025-09-07T07:29:22.7004204Z fi 2025-09-07T07:29:22.7004371Z  2025-09-07T07:29:22.7004674Z PREVIOUS_DOCKER_TAG=$(git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}") 2025-09-07T07:29:22.7005247Z # If no image exists but the hash is the same as the previous hash then we should error out here 2025-09-07T07:29:22.7005756Z if [[ "${PREVIOUS_DOCKER_TAG}" == "${DOCKER_TAG}" ]]; then 2025-09-07T07:29:22.7006349Z  echo "WARNING: Something has gone wrong and the previous image isn't available for the merge-base of your branch" 2025-09-07T07:29:22.7007008Z  echo " Will re-build docker image to store in local cache, TTS may be longer" 2025-09-07T07:29:22.7007391Z fi 2025-09-07T07:29:22.7007576Z  2025-09-07T07:29:22.7007788Z echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-09-07T07:29:22.7013358Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:29:22.7013673Z env: 2025-09-07T07:29:22.7013848Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:22.7014071Z DOCKER_BUILD_DIR: .ci/docker 2025-09-07T07:29:22.7014358Z BASE_REVISION: 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:29:22.7015354Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:22.7016508Z DOCKER_TAG: pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:22.7017176Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:29:22.7017496Z DOCKER_PUSH: 2025-09-07T07:29:22.7017675Z ##[endgroup] 2025-09-07T07:29:22.7039159Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:29:22.7039532Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:29:22.7041436Z + aws ecr get-login-password --region us-east-1 2025-09-07T07:29:22.7042641Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:29:23.1662049Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-09-07T07:29:23.1662577Z Configure a credential helper to remove this warning. See 2025-09-07T07:29:23.1663107Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-09-07T07:29:23.1663445Z 2025-09-07T07:29:23.1663753Z Login Succeeded 2025-09-07T07:29:23.1679091Z ++ date +%s 2025-09-07T07:29:23.1688424Z + START_TIME=1757230163 2025-09-07T07:29:23.1690857Z ++ date +%s 2025-09-07T07:29:23.1699059Z + [[ 1757222963 -lt 1757230163 ]] 2025-09-07T07:29:23.1699920Z + docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:23.3651723Z { 2025-09-07T07:29:23.3651913Z "schemaVersion": 2, 2025-09-07T07:29:23.3652272Z "mediaType": "application/vnd.docker.distribution.manifest.v2+json", 2025-09-07T07:29:23.3652790Z "config": { 2025-09-07T07:29:23.3653073Z "mediaType": "application/vnd.docker.container.image.v1+json", 2025-09-07T07:29:23.3653413Z "size": 30319, 2025-09-07T07:29:23.3653788Z "digest": "sha256:2e41fa7f17d613e4abb96e4b9f7f4a2b2a08134df169ebcbcbc5a819408dc5b3" 2025-09-07T07:29:23.3654193Z }, 2025-09-07T07:29:23.3654354Z "layers": [ 2025-09-07T07:29:23.3654519Z { 2025-09-07T07:29:23.3654800Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3655158Z "size": 28405356, 2025-09-07T07:29:23.3655527Z "digest": "sha256:8bbfefac75f44cec45a58cc5e5ea42260017a8f2e72fa4fa9663869eb9550bc5" 2025-09-07T07:29:23.3655933Z }, 2025-09-07T07:29:23.3656084Z { 2025-09-07T07:29:23.3656367Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3656721Z "size": 1552, 2025-09-07T07:29:23.3657059Z "digest": "sha256:1034e845500b158b573f0409744970fcd7569e41994d2cf1df94c8ac21482591" 2025-09-07T07:29:23.3657447Z }, 2025-09-07T07:29:23.3657597Z { 2025-09-07T07:29:23.3657871Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3658221Z "size": 294587780, 2025-09-07T07:29:23.3658573Z "digest": "sha256:7108634f1291131168bf73f70b53510f8c46e8ff67b0db6a5e8e0e853821358f" 2025-09-07T07:29:23.3658964Z }, 2025-09-07T07:29:23.3659114Z { 2025-09-07T07:29:23.3659387Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3659743Z "size": 793, 2025-09-07T07:29:23.3660090Z "digest": "sha256:9f402ac333db269956ff0962a6fcc21532d1a1193ad3521a0cf0a600db20f158" 2025-09-07T07:29:23.3660484Z }, 2025-09-07T07:29:23.3660633Z { 2025-09-07T07:29:23.3660907Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3661254Z "size": 106, 2025-09-07T07:29:23.3661601Z "digest": "sha256:b2a4ed7220589ec856db4f6cd24774115e4217aa3f9c6e71a7f40ab2793edf0d" 2025-09-07T07:29:23.3661994Z }, 2025-09-07T07:29:23.3662143Z { 2025-09-07T07:29:23.3662416Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3662767Z "size": 703, 2025-09-07T07:29:23.3663113Z "digest": "sha256:d96fa71aa7d0f56e3601b823213c92aba99f3d3d2c71b5c3bbf7814a63c9ddaa" 2025-09-07T07:29:23.3663771Z }, 2025-09-07T07:29:23.3663922Z { 2025-09-07T07:29:23.3664197Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3664541Z "size": 1215, 2025-09-07T07:29:23.3664893Z "digest": "sha256:ab8dd3557a1f259a146b915c43737c9c02a6b3a36a5a919f9c1095adc0a654c1" 2025-09-07T07:29:23.3665474Z }, 2025-09-07T07:29:23.3665625Z { 2025-09-07T07:29:23.3665893Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3666243Z "size": 485, 2025-09-07T07:29:23.3666587Z "digest": "sha256:fd99b0dcee89374b6c7f88e1b31315f3147cd39f439b6d095cf0f33143049e96" 2025-09-07T07:29:23.3666984Z }, 2025-09-07T07:29:23.3667131Z { 2025-09-07T07:29:23.3667404Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3667753Z "size": 110, 2025-09-07T07:29:23.3668099Z "digest": "sha256:f7b4a101181dca2099d756620a88ffcc29d949ea10d4ace59f8d4c8994a35d39" 2025-09-07T07:29:23.3668497Z }, 2025-09-07T07:29:23.3668875Z { 2025-09-07T07:29:23.3669158Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3669508Z "size": 4787, 2025-09-07T07:29:23.3670020Z "digest": "sha256:aae584ec158009e631741b5634c2be6db6c09f7b4f08c8968ee3f5e9c496ba10" 2025-09-07T07:29:23.3670428Z }, 2025-09-07T07:29:23.3670579Z { 2025-09-07T07:29:23.3670853Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3671199Z "size": 1709, 2025-09-07T07:29:23.3671553Z "digest": "sha256:4906db9d12bf51079b260d6e19bf4bd78e0caa2eab549acc9c21ba1a2ce8d9da" 2025-09-07T07:29:23.3671951Z }, 2025-09-07T07:29:23.3672102Z { 2025-09-07T07:29:23.3672371Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3672718Z "size": 724, 2025-09-07T07:29:23.3673049Z "digest": "sha256:cc22251711f7750b06b6949d5624416e900e1b6fd34782e72e980d859512b935" 2025-09-07T07:29:23.3673437Z }, 2025-09-07T07:29:23.3673583Z { 2025-09-07T07:29:23.3673861Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3674211Z "size": 543, 2025-09-07T07:29:23.3674559Z "digest": "sha256:7022473e3fe92a7cc665c4cd41a061810056afe79e1e2f0bee4f15c106704a92" 2025-09-07T07:29:23.3674947Z }, 2025-09-07T07:29:23.3675097Z { 2025-09-07T07:29:23.3675371Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3675723Z "size": 1761129048, 2025-09-07T07:29:23.3676080Z "digest": "sha256:44dfb511788f5da6ad593a2511be4b8376b66c9c2d6b4db39b95b1f2e4ee4a20" 2025-09-07T07:29:23.3676475Z }, 2025-09-07T07:29:23.3676623Z { 2025-09-07T07:29:23.3676898Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3677244Z "size": 32, 2025-09-07T07:29:23.3677590Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:29:23.3677989Z }, 2025-09-07T07:29:23.3678140Z { 2025-09-07T07:29:23.3678415Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3678766Z "size": 380, 2025-09-07T07:29:23.3679109Z "digest": "sha256:20ee7a0428191d54ace17df174653f580b9ad80586d0c15a8fd5e39f45f4453f" 2025-09-07T07:29:23.3679501Z }, 2025-09-07T07:29:23.3679648Z { 2025-09-07T07:29:23.3679918Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3680265Z "size": 235011, 2025-09-07T07:29:23.3680614Z "digest": "sha256:1acbc52cc958308b5647ea6664b955050876d50a9ec5c1fefde3acae59875f65" 2025-09-07T07:29:23.3681004Z }, 2025-09-07T07:29:23.3681154Z { 2025-09-07T07:29:23.3681428Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3681781Z "size": 232, 2025-09-07T07:29:23.3682122Z "digest": "sha256:56a48ab8970272ebb38a7a1b703ce7df174a6e1ed94127b6fc4893e23ce962d7" 2025-09-07T07:29:23.3682517Z }, 2025-09-07T07:29:23.3682667Z { 2025-09-07T07:29:23.3682940Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3683489Z "size": 3403396, 2025-09-07T07:29:23.3683844Z "digest": "sha256:3f2a38e8e42c73d61590b00e4dcbb435b85b79a838294963286ae92b8a79461f" 2025-09-07T07:29:23.3684233Z }, 2025-09-07T07:29:23.3684382Z { 2025-09-07T07:29:23.3684659Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3685011Z "size": 1478, 2025-09-07T07:29:23.3685345Z "digest": "sha256:67060a7526c790925b4b950967803c68a031f427ed6d2765b6a88589276657f1" 2025-09-07T07:29:23.3685722Z }, 2025-09-07T07:29:23.3685867Z { 2025-09-07T07:29:23.3686142Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3686491Z "size": 483, 2025-09-07T07:29:23.3686825Z "digest": "sha256:5044c8e66435628ca02bea372381f0182b5526090c24a34a5aee70c71bd0c411" 2025-09-07T07:29:23.3687205Z }, 2025-09-07T07:29:23.3687354Z { 2025-09-07T07:29:23.3687628Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3687982Z "size": 194, 2025-09-07T07:29:23.3688333Z "digest": "sha256:e8e3b05e8ebfecb472c6caea0a26f0ed0a98e99d3cd1b8813d9ff035626f821f" 2025-09-07T07:29:23.3688742Z }, 2025-09-07T07:29:23.3688893Z { 2025-09-07T07:29:23.3689277Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3689628Z "size": 608, 2025-09-07T07:29:23.3689975Z "digest": "sha256:f1f79ed6e8378e652d498460fce09f2ee3d03a4e1f14c9d088dc9ee87d8332ff" 2025-09-07T07:29:23.3690371Z }, 2025-09-07T07:29:23.3690520Z { 2025-09-07T07:29:23.3690810Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3702783Z "size": 225, 2025-09-07T07:29:23.3703167Z "digest": "sha256:e9247cb8405c7cb3628c221623cc70a75be4978c057f1969216b6065520675f4" 2025-09-07T07:29:23.3703557Z }, 2025-09-07T07:29:23.3703708Z { 2025-09-07T07:29:23.3703986Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3704335Z "size": 828, 2025-09-07T07:29:23.3704692Z "digest": "sha256:9c69a27176e5a6dd7eeeee6c5af22dcdf4d20a80f5601003486666ecc2426299" 2025-09-07T07:29:23.3705092Z }, 2025-09-07T07:29:23.3705240Z { 2025-09-07T07:29:23.3705517Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3705861Z "size": 32, 2025-09-07T07:29:23.3706205Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:29:23.3706604Z }, 2025-09-07T07:29:23.3706749Z { 2025-09-07T07:29:23.3707018Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3707368Z "size": 104, 2025-09-07T07:29:23.3707715Z "digest": "sha256:e7fb26a8f3334384073f3599a5ca6a8f932d1dd5ccf6be2e29162bb535e4406f" 2025-09-07T07:29:23.3708108Z }, 2025-09-07T07:29:23.3708252Z { 2025-09-07T07:29:23.3708521Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3708867Z "size": 1495, 2025-09-07T07:29:23.3709226Z "digest": "sha256:afdfb47fc1ed4ed02ec394a26cce86402ba7091b41adf49c6b7cdcc0d8fa6760" 2025-09-07T07:29:23.3709634Z }, 2025-09-07T07:29:23.3709781Z { 2025-09-07T07:29:23.3710052Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3710404Z "size": 356789661, 2025-09-07T07:29:23.3710756Z "digest": "sha256:488b5af903b5a33e6e89f899e68e9eb85019d1cc59e4652ca251710f75f1a78e" 2025-09-07T07:29:23.3711147Z }, 2025-09-07T07:29:23.3711295Z { 2025-09-07T07:29:23.3711566Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3711909Z "size": 164, 2025-09-07T07:29:23.3712252Z "digest": "sha256:36e30b1c2982e72aac40f44b831cfb9815f20fd1579faaa709b42581043e3211" 2025-09-07T07:29:23.3712642Z }, 2025-09-07T07:29:23.3712787Z { 2025-09-07T07:29:23.3713052Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3713399Z "size": 347, 2025-09-07T07:29:23.3713742Z "digest": "sha256:05ee2b3d005a82a30d77bd991e78a6b4c5f072c732ccdfa91fb50be244b1529a" 2025-09-07T07:29:23.3714311Z }, 2025-09-07T07:29:23.3714456Z { 2025-09-07T07:29:23.3714728Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3715077Z "size": 32, 2025-09-07T07:29:23.3715425Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:29:23.3715818Z }, 2025-09-07T07:29:23.3715965Z { 2025-09-07T07:29:23.3716233Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3716578Z "size": 106, 2025-09-07T07:29:23.3716919Z "digest": "sha256:6ffeb6d1993e174a0afd2ecb287c34b4e8bd6949c946b6350d56641dd5e3b00c" 2025-09-07T07:29:23.3717313Z }, 2025-09-07T07:29:23.3717458Z { 2025-09-07T07:29:23.3717727Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3718071Z "size": 425, 2025-09-07T07:29:23.3718409Z "digest": "sha256:afeef50b450839155e79485329b571fe5b7654ebe4578cf65b190bf37a3d4b2b" 2025-09-07T07:29:23.3718804Z }, 2025-09-07T07:29:23.3718949Z { 2025-09-07T07:29:23.3719215Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3719561Z "size": 19007351, 2025-09-07T07:29:23.3720032Z "digest": "sha256:3fcf510a2172ed3a97876c86d4d193a295153bb52edb43dd0ee2670f8f14975f" 2025-09-07T07:29:23.3720433Z }, 2025-09-07T07:29:23.3720576Z { 2025-09-07T07:29:23.3720847Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3721189Z "size": 108, 2025-09-07T07:29:23.3721528Z "digest": "sha256:d1b4c895928ee8e22236282d2244dbc55a0eadbd2b1821d5bf043c8702f78fec" 2025-09-07T07:29:23.3721914Z }, 2025-09-07T07:29:23.3722058Z { 2025-09-07T07:29:23.3722327Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3722670Z "size": 636, 2025-09-07T07:29:23.3723019Z "digest": "sha256:69ec6a4b91e83e6ddcbef6bfca12aad793ff3021558df535cd82a355c2818832" 2025-09-07T07:29:23.3723420Z }, 2025-09-07T07:29:23.3723572Z { 2025-09-07T07:29:23.3723843Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3724185Z "size": 724, 2025-09-07T07:29:23.3724513Z "digest": "sha256:cc22251711f7750b06b6949d5624416e900e1b6fd34782e72e980d859512b935" 2025-09-07T07:29:23.3724896Z }, 2025-09-07T07:29:23.3725044Z { 2025-09-07T07:29:23.3725311Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3725655Z "size": 148, 2025-09-07T07:29:23.3725987Z "digest": "sha256:102375337d731d7911db444a4b035d1c0196ac5768b87dce506ed043c8f1b3ee" 2025-09-07T07:29:23.3726366Z }, 2025-09-07T07:29:23.3726512Z { 2025-09-07T07:29:23.3726782Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3727126Z "size": 135, 2025-09-07T07:29:23.3727467Z "digest": "sha256:4accd97ef2e20091b2886b4e85d56629434ecb38a53308a53a4d0b49609a7f36" 2025-09-07T07:29:23.3727856Z }, 2025-09-07T07:29:23.3728000Z { 2025-09-07T07:29:23.3728277Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3728621Z "size": 141, 2025-09-07T07:29:23.3728974Z "digest": "sha256:9ee269fa5570cdf5f5e5e41cf9ebb3ee5ed89df8b1229dca7ea5b05563803012" 2025-09-07T07:29:23.3729370Z }, 2025-09-07T07:29:23.3729521Z { 2025-09-07T07:29:23.3729790Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3730139Z "size": 13818466639, 2025-09-07T07:29:23.3730493Z "digest": "sha256:d193fbc9ee8119f23ec55bf4d0d585a8590604108a20cfe418a63b3030802b72" 2025-09-07T07:29:23.3730888Z }, 2025-09-07T07:29:23.3731033Z { 2025-09-07T07:29:23.3731299Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3731645Z "size": 222, 2025-09-07T07:29:23.3731982Z "digest": "sha256:99a2462c8c29855d63392a2bd6dbbd979fb01512a52510c5cd192b3d8c983994" 2025-09-07T07:29:23.3732367Z }, 2025-09-07T07:29:23.3732512Z { 2025-09-07T07:29:23.3732778Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3733263Z "size": 115, 2025-09-07T07:29:23.3733599Z "digest": "sha256:fe1294e18ed093e269c43993d7430aa3f610843964199ae377834997cfaabee1" 2025-09-07T07:29:23.3733981Z }, 2025-09-07T07:29:23.3734124Z { 2025-09-07T07:29:23.3734397Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3734743Z "size": 32, 2025-09-07T07:29:23.3735088Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:29:23.3735478Z }, 2025-09-07T07:29:23.3735623Z { 2025-09-07T07:29:23.3735890Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3736234Z "size": 128, 2025-09-07T07:29:23.3736567Z "digest": "sha256:f379f1cd3010a4d0d9cfb477264d42695f390ee06b7829ea31b46f508821a0d9" 2025-09-07T07:29:23.3736953Z }, 2025-09-07T07:29:23.3737099Z { 2025-09-07T07:29:23.3737367Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3737713Z "size": 778, 2025-09-07T07:29:23.3738055Z "digest": "sha256:6ae93e16ed4e3630a205666a5dad053e0b5ef6424d20ee6cfe1a07453c948f0f" 2025-09-07T07:29:23.3738447Z }, 2025-09-07T07:29:23.3738593Z { 2025-09-07T07:29:23.3738965Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3739320Z "size": 724, 2025-09-07T07:29:23.3739649Z "digest": "sha256:cc22251711f7750b06b6949d5624416e900e1b6fd34782e72e980d859512b935" 2025-09-07T07:29:23.3740026Z }, 2025-09-07T07:29:23.3740168Z { 2025-09-07T07:29:23.3740435Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3740780Z "size": 140, 2025-09-07T07:29:23.3741122Z "digest": "sha256:f995c3109bd824ffff3dab1e70f914d1d48658c9185bf9ff16621b4cb759c154" 2025-09-07T07:29:23.3741511Z }, 2025-09-07T07:29:23.3741657Z { 2025-09-07T07:29:23.3741921Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3742267Z "size": 32, 2025-09-07T07:29:23.3742611Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:29:23.3743006Z }, 2025-09-07T07:29:23.3743151Z { 2025-09-07T07:29:23.3743420Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3743765Z "size": 160, 2025-09-07T07:29:23.3744101Z "digest": "sha256:058384795baca2774d0ced637115ab37947c3b38cf3b478e361ed1c8ecadce06" 2025-09-07T07:29:23.3744486Z }, 2025-09-07T07:29:23.3744632Z { 2025-09-07T07:29:23.3744899Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3745243Z "size": 1010, 2025-09-07T07:29:23.3745590Z "digest": "sha256:3e762f827dd11a0de09efae2ae65697870981c46a36af548b1245ba5892cde17" 2025-09-07T07:29:23.3745977Z }, 2025-09-07T07:29:23.3746121Z { 2025-09-07T07:29:23.3746388Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3746736Z "size": 724, 2025-09-07T07:29:23.3747066Z "digest": "sha256:cc22251711f7750b06b6949d5624416e900e1b6fd34782e72e980d859512b935" 2025-09-07T07:29:23.3747447Z }, 2025-09-07T07:29:23.3747593Z { 2025-09-07T07:29:23.3747864Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3748207Z "size": 134, 2025-09-07T07:29:23.3748553Z "digest": "sha256:6acfbf2cdc1758ddc14df6ba92219671f9e493ef324754be7400e6ec97f67c39" 2025-09-07T07:29:23.3748950Z }, 2025-09-07T07:29:23.3749097Z { 2025-09-07T07:29:23.3749365Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3749708Z "size": 32, 2025-09-07T07:29:23.3750050Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:29:23.3750449Z }, 2025-09-07T07:29:23.3750597Z { 2025-09-07T07:29:23.3750864Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3751210Z "size": 158, 2025-09-07T07:29:23.3751551Z "digest": "sha256:1e43a43b4f066f3d26e6ef5cb34fe1c729f7ebb72976d92b02367752a5c72bb4" 2025-09-07T07:29:23.3752079Z }, 2025-09-07T07:29:23.3752224Z { 2025-09-07T07:29:23.3752491Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3752836Z "size": 1369, 2025-09-07T07:29:23.3753170Z "digest": "sha256:968c538d5104090ee324d3f7e11d3524e8d282ac51a8187546f9348d705f7391" 2025-09-07T07:29:23.3753545Z }, 2025-09-07T07:29:23.3753692Z { 2025-09-07T07:29:23.3753960Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3754302Z "size": 32, 2025-09-07T07:29:23.3754642Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:29:23.3755034Z }, 2025-09-07T07:29:23.3755180Z { 2025-09-07T07:29:23.3755446Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3755786Z "size": 136, 2025-09-07T07:29:23.3756125Z "digest": "sha256:575807ea3b5f8afba79759ff01fb50da569fa8263726404df19d016f0562bbde" 2025-09-07T07:29:23.3756514Z }, 2025-09-07T07:29:23.3756667Z { 2025-09-07T07:29:23.3756931Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3757276Z "size": 380, 2025-09-07T07:29:23.3757718Z "digest": "sha256:6a61bbb320049660d19a61d909d95a4561774698fc6db2770885aafee600a360" 2025-09-07T07:29:23.3758113Z }, 2025-09-07T07:29:23.3758257Z { 2025-09-07T07:29:23.3758526Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3758879Z "size": 979283315, 2025-09-07T07:29:23.3759245Z "digest": "sha256:f88dbb5c90de18bfddff326da203a7af18680486c3be8c12948926bf9d0ffa79" 2025-09-07T07:29:23.3759647Z }, 2025-09-07T07:29:23.3759795Z { 2025-09-07T07:29:23.3760067Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3760413Z "size": 104, 2025-09-07T07:29:23.3760749Z "digest": "sha256:08443826e2c53c5cac1c198c5db0e7c44c5bc3cc7c644e2ac5c8f96c40566501" 2025-09-07T07:29:23.3761141Z }, 2025-09-07T07:29:23.3761290Z { 2025-09-07T07:29:23.3761565Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3761911Z "size": 407, 2025-09-07T07:29:23.3762252Z "digest": "sha256:923b78b302940958a199d259e5ad1e2d469ddd34aa0e99cfed1eb74e2584fba0" 2025-09-07T07:29:23.3762643Z }, 2025-09-07T07:29:23.3762796Z { 2025-09-07T07:29:23.3763062Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3763409Z "size": 177786912, 2025-09-07T07:29:23.3763766Z "digest": "sha256:0751aba37e5a50eb07a627d73c152e5e49f9868fe35912077806fe3b676d6e82" 2025-09-07T07:29:23.3764154Z }, 2025-09-07T07:29:23.3764300Z { 2025-09-07T07:29:23.3764573Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3764920Z "size": 109, 2025-09-07T07:29:23.3765257Z "digest": "sha256:fbe315c5491d2d86350697f89fea7f2a1450c426f5d3e50f314fa6451543dba8" 2025-09-07T07:29:23.3765738Z }, 2025-09-07T07:29:23.3765885Z { 2025-09-07T07:29:23.3766156Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3766508Z "size": 1896, 2025-09-07T07:29:23.3766846Z "digest": "sha256:af69784d4cfc8f22cd48487f7a473b334683969a216ae75357a59a6bd161fbc0" 2025-09-07T07:29:23.3767235Z }, 2025-09-07T07:29:23.3767384Z { 2025-09-07T07:29:23.3767661Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3768006Z "size": 236755277, 2025-09-07T07:29:23.3768363Z "digest": "sha256:91e4deab4f89beb8fd430738fe6ae39256401ed1bf3b256a566c6265fab2dca6" 2025-09-07T07:29:23.3768761Z }, 2025-09-07T07:29:23.3768909Z { 2025-09-07T07:29:23.3769176Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3769524Z "size": 106, 2025-09-07T07:29:23.3769860Z "digest": "sha256:c1d56a01ea406308c34e9089673abb778aede0510d9280b45e1a0151a4343c10" 2025-09-07T07:29:23.3770247Z }, 2025-09-07T07:29:23.3770392Z { 2025-09-07T07:29:23.3770668Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3771223Z "size": 163, 2025-09-07T07:29:23.3771572Z "digest": "sha256:9ad931d990d437134cd1401c20794b6f851d023afa54fbce4166a34cdf599544" 2025-09-07T07:29:23.3771958Z }, 2025-09-07T07:29:23.3772106Z { 2025-09-07T07:29:23.3772382Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3772732Z "size": 7943, 2025-09-07T07:29:23.3773080Z "digest": "sha256:df5b3d079e9e5cdd3d2e21d2fdc4ffce8b71ca9940e685b1065f99b486762ed4" 2025-09-07T07:29:23.3773483Z }, 2025-09-07T07:29:23.3773631Z { 2025-09-07T07:29:23.3773901Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3774246Z "size": 8075, 2025-09-07T07:29:23.3774579Z "digest": "sha256:24f5a79527e36e783e09941f86d47277429c78c36137b0f1ca5b5599080894dd" 2025-09-07T07:29:23.3774962Z }, 2025-09-07T07:29:23.3775110Z { 2025-09-07T07:29:23.3775376Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3775723Z "size": 302, 2025-09-07T07:29:23.3776069Z "digest": "sha256:4166cbf927c324e2961271c43f4fd4604caab5bcb537e5e3cbe3a89599f5caef" 2025-09-07T07:29:23.3776463Z }, 2025-09-07T07:29:23.3776610Z { 2025-09-07T07:29:23.3777058Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3777419Z "size": 32, 2025-09-07T07:29:23.3777768Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:29:23.3778161Z }, 2025-09-07T07:29:23.3778310Z { 2025-09-07T07:29:23.3778580Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3778926Z "size": 108, 2025-09-07T07:29:23.3779253Z "digest": "sha256:d243494d377d03f282a3e8703417520396fa042420dabd7730848f55db3969a6" 2025-09-07T07:29:23.3779633Z }, 2025-09-07T07:29:23.3779780Z { 2025-09-07T07:29:23.3780050Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3780397Z "size": 54145658, 2025-09-07T07:29:23.3780752Z "digest": "sha256:8a28267fa33261af473c2c4f617903a0c7af7cf2203014da2f37b5dd58aac207" 2025-09-07T07:29:23.3781142Z }, 2025-09-07T07:29:23.3781289Z { 2025-09-07T07:29:23.3781557Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:29:23.3781909Z "size": 125, 2025-09-07T07:29:23.3782261Z "digest": "sha256:89735be27bd4dc00abc5df9a6483a79fe2dd2df2e86d6fa7db0fc871b1a6df57" 2025-09-07T07:29:23.3782663Z } 2025-09-07T07:29:23.3782807Z ] 2025-09-07T07:29:23.3782955Z } 2025-09-07T07:29:23.3783121Z + exit 0 2025-09-07T07:29:23.3807867Z ##[group]Run set -eux 2025-09-07T07:29:23.3808098Z set -eux 2025-09-07T07:29:23.3808460Z # It's ok if this steps fails, it would then be an anonymous user like what we used to have 2025-09-07T07:29:23.3809458Z aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token | jq --raw-output '.SecretString' | jq -r .docker_hub_readonly_token | docker login --username pytorchbot --password-stdin || true 2025-09-07T07:29:23.3817043Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:29:23.3817360Z env: 2025-09-07T07:29:23.3817534Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:23.3817746Z ##[endgroup] 2025-09-07T07:29:23.3844303Z + aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token 2025-09-07T07:29:23.3845012Z + jq --raw-output .SecretString 2025-09-07T07:29:23.3846009Z + jq -r .docker_hub_readonly_token 2025-09-07T07:29:23.3846749Z + docker login --username pytorchbot --password-stdin 2025-09-07T07:29:23.9060375Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-09-07T07:29:23.9060893Z Configure a credential helper to remove this warning. See 2025-09-07T07:29:23.9061390Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-09-07T07:29:23.9061726Z 2025-09-07T07:29:23.9062101Z Login Succeeded 2025-09-07T07:29:23.9147194Z ##[group]Run tag=${ECR_DOCKER_IMAGE##*:} 2025-09-07T07:29:23.9147581Z tag=${ECR_DOCKER_IMAGE##*:} 2025-09-07T07:29:23.9148134Z echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}" 2025-09-07T07:29:23.9154753Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:29:23.9155066Z env: 2025-09-07T07:29:23.9155238Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:23.9156042Z ECR_DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:23.9156855Z ##[endgroup] 2025-09-07T07:29:23.9181529Z docker pull ghcr.io/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:23.9221543Z ##[group]Run pytorch/test-infra/.github/actions/pull-docker-image@main 2025-09-07T07:29:23.9221938Z with: 2025-09-07T07:29:23.9222691Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:23.9223610Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:29:23.9223926Z env: 2025-09-07T07:29:23.9224097Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:23.9224309Z ##[endgroup] 2025-09-07T07:29:23.9237514Z ##[group]Run set -x 2025-09-07T07:29:23.9237739Z set -x 2025-09-07T07:29:23.9237921Z set +e 2025-09-07T07:29:23.9238097Z  2025-09-07T07:29:23.9238258Z login() { 2025-09-07T07:29:23.9238664Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-09-07T07:29:23.9239105Z } 2025-09-07T07:29:23.9239270Z  2025-09-07T07:29:23.9239463Z retry () { 2025-09-07T07:29:23.9239693Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-09-07T07:29:23.9239955Z } 2025-09-07T07:29:23.9240121Z  2025-09-07T07:29:23.9240313Z retry login "${DOCKER_REGISTRY}" 2025-09-07T07:29:23.9240576Z  2025-09-07T07:29:23.9240986Z IMAGE_SIZE=$(docker manifest inspect "${DOCKER_IMAGE}" | jq '[.layers[].size, .config.size] | add / 1024 / 1024') 2025-09-07T07:29:23.9241550Z echo "Compressed size of image in MB: ${IMAGE_SIZE}" 2025-09-07T07:29:23.9241861Z  2025-09-07T07:29:23.9242027Z set -e 2025-09-07T07:29:23.9242310Z # ignore output since only exit code is used for conditional 2025-09-07T07:29:23.9242729Z # only pull docker image if it's not available locally 2025-09-07T07:29:23.9243192Z if ! docker inspect --type=image "${DOCKER_IMAGE}" >/dev/null 2>/dev/null; then 2025-09-07T07:29:23.9243630Z  retry docker pull "${DOCKER_IMAGE}" 2025-09-07T07:29:23.9243898Z fi 2025-09-07T07:29:23.9249502Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:29:23.9249820Z env: 2025-09-07T07:29:23.9249996Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:29:23.9250791Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:23.9251699Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:29:23.9252022Z ##[endgroup] 2025-09-07T07:29:23.9272549Z + set +e 2025-09-07T07:29:23.9272881Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:29:23.9273253Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:29:23.9275670Z + aws ecr get-login-password --region us-east-1 2025-09-07T07:29:23.9276466Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:29:24.3949314Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-09-07T07:29:24.3949839Z Configure a credential helper to remove this warning. See 2025-09-07T07:29:24.3950331Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-09-07T07:29:24.3950976Z 2025-09-07T07:29:24.3951130Z Login Succeeded 2025-09-07T07:29:24.3972171Z ++ docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:24.3973094Z ++ jq '[.layers[].size, .config.size] | add / 1024 / 1024' 2025-09-07T07:29:24.6186905Z + IMAGE_SIZE=16908.71542263031 2025-09-07T07:29:24.6187215Z + echo 'Compressed size of image in MB: 16908.71542263031' 2025-09-07T07:29:24.6187519Z + set -e 2025-09-07T07:29:24.6187735Z Compressed size of image in MB: 16908.71542263031 2025-09-07T07:29:24.6188924Z + docker inspect --type=image 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:24.6318144Z + retry docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:24.6319565Z + docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:29:24.8306583Z pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77: Pulling from pytorch/ci-image 2025-09-07T07:29:24.8307578Z 8bbfefac75f4: Pulling fs layer 2025-09-07T07:29:24.8307852Z 1034e845500b: Pulling fs layer 2025-09-07T07:29:24.8308083Z 7108634f1291: Pulling fs layer 2025-09-07T07:29:24.8308329Z 9f402ac333db: Pulling fs layer 2025-09-07T07:29:24.8308559Z b2a4ed722058: Pulling fs layer 2025-09-07T07:29:24.8308810Z d96fa71aa7d0: Pulling fs layer 2025-09-07T07:29:24.8309051Z ab8dd3557a1f: Pulling fs layer 2025-09-07T07:29:24.8309284Z fd99b0dcee89: Pulling fs layer 2025-09-07T07:29:24.8309513Z f7b4a101181d: Pulling fs layer 2025-09-07T07:29:24.8309744Z aae584ec1580: Pulling fs layer 2025-09-07T07:29:24.8309991Z 4906db9d12bf: Pulling fs layer 2025-09-07T07:29:24.8310222Z cc22251711f7: Pulling fs layer 2025-09-07T07:29:24.8310450Z 7022473e3fe9: Pulling fs layer 2025-09-07T07:29:24.8310679Z 44dfb511788f: Pulling fs layer 2025-09-07T07:29:24.8310911Z 4f4fb700ef54: Pulling fs layer 2025-09-07T07:29:24.8311141Z 20ee7a042819: Pulling fs layer 2025-09-07T07:29:24.8311371Z 1acbc52cc958: Pulling fs layer 2025-09-07T07:29:24.8311600Z 56a48ab89702: Pulling fs layer 2025-09-07T07:29:24.8311829Z 3f2a38e8e42c: Pulling fs layer 2025-09-07T07:29:24.8312058Z 67060a7526c7: Pulling fs layer 2025-09-07T07:29:24.8312282Z 5044c8e66435: Pulling fs layer 2025-09-07T07:29:24.8312557Z b2a4ed722058: Waiting 2025-09-07T07:29:24.8312778Z e8e3b05e8ebf: Pulling fs layer 2025-09-07T07:29:24.8313014Z f1f79ed6e837: Pulling fs layer 2025-09-07T07:29:24.8313234Z d96fa71aa7d0: Waiting 2025-09-07T07:29:24.8313438Z e9247cb8405c: Pulling fs layer 2025-09-07T07:29:24.8313667Z 9c69a27176e5: Pulling fs layer 2025-09-07T07:29:24.8313894Z f7b4a101181d: Waiting 2025-09-07T07:29:24.8314088Z aae584ec1580: Waiting 2025-09-07T07:29:24.8314278Z ab8dd3557a1f: Waiting 2025-09-07T07:29:24.8314484Z e7fb26a8f333: Pulling fs layer 2025-09-07T07:29:24.8314723Z afdfb47fc1ed: Pulling fs layer 2025-09-07T07:29:24.8314955Z 488b5af903b5: Pulling fs layer 2025-09-07T07:29:24.8315180Z 36e30b1c2982: Pulling fs layer 2025-09-07T07:29:24.8315410Z 05ee2b3d005a: Pulling fs layer 2025-09-07T07:29:24.8315627Z 4906db9d12bf: Waiting 2025-09-07T07:29:24.8315821Z fd99b0dcee89: Waiting 2025-09-07T07:29:24.8316008Z 7022473e3fe9: Waiting 2025-09-07T07:29:24.8316197Z cc22251711f7: Waiting 2025-09-07T07:29:24.8316398Z 6ffeb6d1993e: Pulling fs layer 2025-09-07T07:29:24.8316641Z afeef50b4508: Pulling fs layer 2025-09-07T07:29:24.8316860Z 44dfb511788f: Waiting 2025-09-07T07:29:24.8317051Z 56a48ab89702: Waiting 2025-09-07T07:29:24.8317240Z 1acbc52cc958: Waiting 2025-09-07T07:29:24.8317441Z 3fcf510a2172: Pulling fs layer 2025-09-07T07:29:24.8317668Z d1b4c895928e: Pulling fs layer 2025-09-07T07:29:24.8318196Z 67060a7526c7: Waiting 2025-09-07T07:29:24.8318400Z 69ec6a4b91e8: Pulling fs layer 2025-09-07T07:29:24.8318620Z f1f79ed6e837: Waiting 2025-09-07T07:29:24.8318811Z afdfb47fc1ed: Waiting 2025-09-07T07:29:24.8319014Z 102375337d73: Pulling fs layer 2025-09-07T07:29:24.8319226Z 9c69a27176e5: Waiting 2025-09-07T07:29:24.8329383Z 3f2a38e8e42c: Waiting 2025-09-07T07:29:24.8329609Z 20ee7a042819: Waiting 2025-09-07T07:29:24.8329822Z 4accd97ef2e2: Pulling fs layer 2025-09-07T07:29:24.8330049Z 9f402ac333db: Waiting 2025-09-07T07:29:24.8330241Z e7fb26a8f333: Waiting 2025-09-07T07:29:24.8330434Z e9247cb8405c: Waiting 2025-09-07T07:29:24.8330632Z 9ee269fa5570: Pulling fs layer 2025-09-07T07:29:24.8331077Z 4f4fb700ef54: Waiting 2025-09-07T07:29:24.8331280Z 5044c8e66435: Waiting 2025-09-07T07:29:24.8331487Z d193fbc9ee81: Pulling fs layer 2025-09-07T07:29:24.8331708Z e8e3b05e8ebf: Waiting 2025-09-07T07:29:24.8331913Z 99a2462c8c29: Pulling fs layer 2025-09-07T07:29:24.8332131Z 488b5af903b5: Waiting 2025-09-07T07:29:24.8332329Z 3fcf510a2172: Waiting 2025-09-07T07:29:24.8332528Z fe1294e18ed0: Pulling fs layer 2025-09-07T07:29:24.8332751Z 05ee2b3d005a: Waiting 2025-09-07T07:29:24.8332940Z 36e30b1c2982: Waiting 2025-09-07T07:29:24.8333129Z 69ec6a4b91e8: Waiting 2025-09-07T07:29:24.8333316Z 6ffeb6d1993e: Waiting 2025-09-07T07:29:24.8333505Z 102375337d73: Waiting 2025-09-07T07:29:24.8333704Z f379f1cd3010: Pulling fs layer 2025-09-07T07:29:24.8333924Z 9ee269fa5570: Waiting 2025-09-07T07:29:24.8334108Z 99a2462c8c29: Waiting 2025-09-07T07:29:24.8334298Z 4accd97ef2e2: Waiting 2025-09-07T07:29:24.8334503Z 6ae93e16ed4e: Pulling fs layer 2025-09-07T07:29:24.8334725Z afeef50b4508: Waiting 2025-09-07T07:29:24.8334919Z fe1294e18ed0: Waiting 2025-09-07T07:29:24.8335113Z d193fbc9ee81: Waiting 2025-09-07T07:29:24.8335315Z f995c3109bd8: Pulling fs layer 2025-09-07T07:29:24.8335536Z 6ae93e16ed4e: Waiting 2025-09-07T07:29:24.8335738Z 058384795bac: Pulling fs layer 2025-09-07T07:29:24.8335966Z 3e762f827dd1: Pulling fs layer 2025-09-07T07:29:24.8336206Z 6acfbf2cdc17: Pulling fs layer 2025-09-07T07:29:24.8336443Z 1e43a43b4f06: Pulling fs layer 2025-09-07T07:29:24.8336668Z 968c538d5104: Pulling fs layer 2025-09-07T07:29:24.8336881Z 1e43a43b4f06: Waiting 2025-09-07T07:29:24.8337084Z 575807ea3b5f: Pulling fs layer 2025-09-07T07:29:24.8337298Z 968c538d5104: Waiting 2025-09-07T07:29:24.8337489Z 058384795bac: Waiting 2025-09-07T07:29:24.8337682Z 3e762f827dd1: Waiting 2025-09-07T07:29:24.8337872Z f995c3109bd8: Waiting 2025-09-07T07:29:24.8338071Z 6a61bbb32004: Pulling fs layer 2025-09-07T07:29:24.8338291Z 575807ea3b5f: Waiting 2025-09-07T07:29:24.8338496Z f88dbb5c90de: Pulling fs layer 2025-09-07T07:29:24.8338715Z 6a61bbb32004: Waiting 2025-09-07T07:29:24.8338919Z 08443826e2c5: Pulling fs layer 2025-09-07T07:29:24.8339149Z 923b78b30294: Pulling fs layer 2025-09-07T07:29:24.8339377Z 0751aba37e5a: Pulling fs layer 2025-09-07T07:29:24.8339610Z fbe315c5491d: Pulling fs layer 2025-09-07T07:29:24.8339836Z af69784d4cfc: Pulling fs layer 2025-09-07T07:29:24.8340073Z 91e4deab4f89: Pulling fs layer 2025-09-07T07:29:24.8340295Z f88dbb5c90de: Waiting 2025-09-07T07:29:24.8340490Z 0751aba37e5a: Waiting 2025-09-07T07:29:24.8340691Z c1d56a01ea40: Pulling fs layer 2025-09-07T07:29:24.8340912Z fbe315c5491d: Waiting 2025-09-07T07:29:24.8341127Z 91e4deab4f89: Waiting 2025-09-07T07:29:24.8341317Z 08443826e2c5: Waiting 2025-09-07T07:29:24.8341515Z 9ad931d990d4: Pulling fs layer 2025-09-07T07:29:24.8341733Z c1d56a01ea40: Waiting 2025-09-07T07:29:24.8341939Z df5b3d079e9e: Pulling fs layer 2025-09-07T07:29:24.8342170Z 24f5a79527e3: Pulling fs layer 2025-09-07T07:29:24.8342393Z 4166cbf927c3: Pulling fs layer 2025-09-07T07:29:24.8342623Z d243494d377d: Pulling fs layer 2025-09-07T07:29:24.8342854Z 8a28267fa332: Pulling fs layer 2025-09-07T07:29:24.8343080Z 89735be27bd4: Pulling fs layer 2025-09-07T07:29:24.8343291Z 9ad931d990d4: Waiting 2025-09-07T07:29:24.8343485Z df5b3d079e9e: Waiting 2025-09-07T07:29:24.8343678Z 89735be27bd4: Waiting 2025-09-07T07:29:24.8343868Z 24f5a79527e3: Waiting 2025-09-07T07:29:24.8344206Z 4166cbf927c3: Waiting 2025-09-07T07:29:24.8344392Z 8a28267fa332: Waiting 2025-09-07T07:29:24.8344630Z d243494d377d: Waiting 2025-09-07T07:29:24.9041003Z 1034e845500b: Verifying Checksum 2025-09-07T07:29:24.9041255Z 1034e845500b: Download complete 2025-09-07T07:29:24.9705864Z 9f402ac333db: Verifying Checksum 2025-09-07T07:29:24.9706128Z 9f402ac333db: Download complete 2025-09-07T07:29:25.0396850Z b2a4ed722058: Download complete 2025-09-07T07:29:25.1251112Z d96fa71aa7d0: Verifying Checksum 2025-09-07T07:29:25.1251387Z d96fa71aa7d0: Download complete 2025-09-07T07:29:25.1593944Z 8bbfefac75f4: Verifying Checksum 2025-09-07T07:29:25.1594207Z 8bbfefac75f4: Download complete 2025-09-07T07:29:25.2096061Z ab8dd3557a1f: Verifying Checksum 2025-09-07T07:29:25.2096365Z ab8dd3557a1f: Download complete 2025-09-07T07:29:25.2332070Z fd99b0dcee89: Verifying Checksum 2025-09-07T07:29:25.2332322Z fd99b0dcee89: Download complete 2025-09-07T07:29:25.2726958Z f7b4a101181d: Download complete 2025-09-07T07:29:25.3040505Z aae584ec1580: Download complete 2025-09-07T07:29:25.3408238Z 4906db9d12bf: Verifying Checksum 2025-09-07T07:29:25.3408490Z 4906db9d12bf: Download complete 2025-09-07T07:29:25.3692396Z cc22251711f7: Verifying Checksum 2025-09-07T07:29:25.3692987Z cc22251711f7: Download complete 2025-09-07T07:29:25.4080679Z 7022473e3fe9: Download complete 2025-09-07T07:29:25.4159637Z 4f4fb700ef54: Verifying Checksum 2025-09-07T07:29:25.4159880Z 4f4fb700ef54: Download complete 2025-09-07T07:29:25.4838979Z 20ee7a042819: Verifying Checksum 2025-09-07T07:29:25.4839234Z 20ee7a042819: Download complete 2025-09-07T07:29:25.5562760Z 1acbc52cc958: Verifying Checksum 2025-09-07T07:29:25.5563060Z 1acbc52cc958: Download complete 2025-09-07T07:29:25.6227079Z 56a48ab89702: Verifying Checksum 2025-09-07T07:29:25.6227344Z 56a48ab89702: Download complete 2025-09-07T07:29:25.7106371Z 8bbfefac75f4: Pull complete 2025-09-07T07:29:25.7223544Z 3f2a38e8e42c: Verifying Checksum 2025-09-07T07:29:25.7223811Z 3f2a38e8e42c: Download complete 2025-09-07T07:29:25.7315684Z 1034e845500b: Pull complete 2025-09-07T07:29:25.8088091Z 67060a7526c7: Verifying Checksum 2025-09-07T07:29:25.8088403Z 67060a7526c7: Download complete 2025-09-07T07:29:25.8721773Z 5044c8e66435: Verifying Checksum 2025-09-07T07:29:25.8722049Z 5044c8e66435: Download complete 2025-09-07T07:29:25.9381538Z e8e3b05e8ebf: Verifying Checksum 2025-09-07T07:29:25.9381794Z e8e3b05e8ebf: Download complete 2025-09-07T07:29:26.0192015Z f1f79ed6e837: Verifying Checksum 2025-09-07T07:29:26.0192287Z f1f79ed6e837: Download complete 2025-09-07T07:29:26.1155307Z e9247cb8405c: Verifying Checksum 2025-09-07T07:29:26.1155580Z e9247cb8405c: Download complete 2025-09-07T07:29:26.1820901Z 9c69a27176e5: Verifying Checksum 2025-09-07T07:29:26.1821150Z 9c69a27176e5: Download complete 2025-09-07T07:29:26.2588033Z e7fb26a8f333: Download complete 2025-09-07T07:29:26.3563231Z afdfb47fc1ed: Verifying Checksum 2025-09-07T07:29:26.3563485Z afdfb47fc1ed: Download complete 2025-09-07T07:29:27.8398400Z 7108634f1291: Verifying Checksum 2025-09-07T07:29:27.8398714Z 7108634f1291: Download complete 2025-09-07T07:29:27.9019089Z 36e30b1c2982: Verifying Checksum 2025-09-07T07:29:27.9019388Z 36e30b1c2982: Download complete 2025-09-07T07:29:27.9712801Z 05ee2b3d005a: Verifying Checksum 2025-09-07T07:29:27.9713077Z 05ee2b3d005a: Download complete 2025-09-07T07:29:28.0313331Z 6ffeb6d1993e: Verifying Checksum 2025-09-07T07:29:28.0313591Z 6ffeb6d1993e: Download complete 2025-09-07T07:29:28.1178327Z afeef50b4508: Verifying Checksum 2025-09-07T07:29:28.1178592Z afeef50b4508: Download complete 2025-09-07T07:29:28.3514965Z 3fcf510a2172: Verifying Checksum 2025-09-07T07:29:28.3515275Z 3fcf510a2172: Download complete 2025-09-07T07:29:28.4267617Z d1b4c895928e: Download complete 2025-09-07T07:29:28.4920394Z 69ec6a4b91e8: Download complete 2025-09-07T07:29:28.5525308Z 102375337d73: Verifying Checksum 2025-09-07T07:29:28.5525578Z 102375337d73: Download complete 2025-09-07T07:29:28.6375448Z 4accd97ef2e2: Verifying Checksum 2025-09-07T07:29:28.6376148Z 4accd97ef2e2: Download complete 2025-09-07T07:29:28.7249137Z 9ee269fa5570: Verifying Checksum 2025-09-07T07:29:28.7249427Z 9ee269fa5570: Download complete 2025-09-07T07:29:29.9632355Z 488b5af903b5: Verifying Checksum 2025-09-07T07:29:29.9632668Z 488b5af903b5: Download complete 2025-09-07T07:29:30.0382198Z 99a2462c8c29: Verifying Checksum 2025-09-07T07:29:30.0382448Z 99a2462c8c29: Download complete 2025-09-07T07:29:30.1122489Z fe1294e18ed0: Verifying Checksum 2025-09-07T07:29:30.1122785Z fe1294e18ed0: Download complete 2025-09-07T07:29:30.1683851Z f379f1cd3010: Verifying Checksum 2025-09-07T07:29:30.1684152Z f379f1cd3010: Download complete 2025-09-07T07:29:30.2516095Z 6ae93e16ed4e: Download complete 2025-09-07T07:29:30.3245792Z f995c3109bd8: Download complete 2025-09-07T07:29:30.3986522Z 058384795bac: Download complete 2025-09-07T07:29:30.4936060Z 3e762f827dd1: Verifying Checksum 2025-09-07T07:29:30.4936416Z 3e762f827dd1: Download complete 2025-09-07T07:29:30.5714924Z 6acfbf2cdc17: Verifying Checksum 2025-09-07T07:29:30.5715248Z 6acfbf2cdc17: Download complete 2025-09-07T07:29:30.6611693Z 1e43a43b4f06: Download complete 2025-09-07T07:29:30.7417907Z 968c538d5104: Verifying Checksum 2025-09-07T07:29:30.7418183Z 968c538d5104: Download complete 2025-09-07T07:29:30.8041081Z 575807ea3b5f: Verifying Checksum 2025-09-07T07:29:30.8041360Z 575807ea3b5f: Download complete 2025-09-07T07:29:30.8596517Z 6a61bbb32004: Verifying Checksum 2025-09-07T07:29:30.8596922Z 6a61bbb32004: Download complete 2025-09-07T07:29:32.5988781Z 7108634f1291: Pull complete 2025-09-07T07:29:32.6239548Z 9f402ac333db: Pull complete 2025-09-07T07:29:32.6467887Z b2a4ed722058: Pull complete 2025-09-07T07:29:32.6707302Z d96fa71aa7d0: Pull complete 2025-09-07T07:29:32.6966910Z ab8dd3557a1f: Pull complete 2025-09-07T07:29:32.7193755Z fd99b0dcee89: Pull complete 2025-09-07T07:29:32.7418823Z f7b4a101181d: Pull complete 2025-09-07T07:29:32.7636651Z aae584ec1580: Pull complete 2025-09-07T07:29:32.7877411Z 4906db9d12bf: Pull complete 2025-09-07T07:29:32.8075743Z cc22251711f7: Pull complete 2025-09-07T07:29:32.8335183Z 7022473e3fe9: Pull complete 2025-09-07T07:29:40.6842756Z f88dbb5c90de: Verifying Checksum 2025-09-07T07:29:40.6843077Z f88dbb5c90de: Download complete 2025-09-07T07:29:40.7561258Z 08443826e2c5: Verifying Checksum 2025-09-07T07:29:40.7561517Z 08443826e2c5: Download complete 2025-09-07T07:29:40.8406976Z 923b78b30294: Verifying Checksum 2025-09-07T07:29:40.8407228Z 923b78b30294: Download complete 2025-09-07T07:29:42.6647919Z 0751aba37e5a: Verifying Checksum 2025-09-07T07:29:42.6648200Z 0751aba37e5a: Download complete 2025-09-07T07:29:42.7505086Z fbe315c5491d: Verifying Checksum 2025-09-07T07:29:42.7505349Z fbe315c5491d: Download complete 2025-09-07T07:29:42.8288128Z af69784d4cfc: Verifying Checksum 2025-09-07T07:29:42.8288384Z af69784d4cfc: Download complete 2025-09-07T07:29:43.0192646Z 44dfb511788f: Verifying Checksum 2025-09-07T07:29:43.0192915Z 44dfb511788f: Download complete 2025-09-07T07:29:43.0867499Z c1d56a01ea40: Download complete 2025-09-07T07:29:43.1636673Z 9ad931d990d4: Download complete 2025-09-07T07:29:43.2503222Z df5b3d079e9e: Verifying Checksum 2025-09-07T07:29:43.2503470Z df5b3d079e9e: Download complete 2025-09-07T07:29:43.3173908Z 24f5a79527e3: Download complete 2025-09-07T07:29:43.3949122Z 4166cbf927c3: Download complete 2025-09-07T07:29:43.4794328Z d243494d377d: Verifying Checksum 2025-09-07T07:29:43.4794614Z d243494d377d: Download complete 2025-09-07T07:29:44.0612134Z 8a28267fa332: Verifying Checksum 2025-09-07T07:29:44.0612406Z 8a28267fa332: Download complete 2025-09-07T07:29:44.1478199Z 89735be27bd4: Verifying Checksum 2025-09-07T07:29:44.1478484Z 89735be27bd4: Download complete 2025-09-07T07:29:45.2610647Z 91e4deab4f89: Verifying Checksum 2025-09-07T07:29:45.2610927Z 91e4deab4f89: Download complete 2025-09-07T07:30:04.0828966Z 44dfb511788f: Pull complete 2025-09-07T07:30:04.1052110Z 4f4fb700ef54: Pull complete 2025-09-07T07:30:04.1635629Z 20ee7a042819: Pull complete 2025-09-07T07:30:04.3334635Z 1acbc52cc958: Pull complete 2025-09-07T07:30:04.7593499Z 56a48ab89702: Pull complete 2025-09-07T07:30:05.1805792Z 3f2a38e8e42c: Pull complete 2025-09-07T07:30:05.5105398Z 67060a7526c7: Pull complete 2025-09-07T07:30:05.9194504Z 5044c8e66435: Pull complete 2025-09-07T07:30:06.2608914Z e8e3b05e8ebf: Pull complete 2025-09-07T07:30:06.4854364Z f1f79ed6e837: Pull complete 2025-09-07T07:30:06.6728292Z e9247cb8405c: Pull complete 2025-09-07T07:30:06.9503218Z 9c69a27176e5: Pull complete 2025-09-07T07:30:07.7618860Z e7fb26a8f333: Pull complete 2025-09-07T07:30:08.1040724Z afdfb47fc1ed: Pull complete 2025-09-07T07:30:13.7080058Z 488b5af903b5: Pull complete 2025-09-07T07:30:13.7739801Z 36e30b1c2982: Pull complete 2025-09-07T07:30:13.9020851Z 05ee2b3d005a: Pull complete 2025-09-07T07:30:14.2721050Z 6ffeb6d1993e: Pull complete 2025-09-07T07:30:14.3560218Z afeef50b4508: Pull complete 2025-09-07T07:30:14.7423397Z 3fcf510a2172: Pull complete 2025-09-07T07:30:15.1918989Z d1b4c895928e: Pull complete 2025-09-07T07:30:15.6293778Z 69ec6a4b91e8: Pull complete 2025-09-07T07:30:16.3967273Z 102375337d73: Pull complete 2025-09-07T07:30:16.8551981Z 4accd97ef2e2: Pull complete 2025-09-07T07:30:17.2433505Z 9ee269fa5570: Pull complete 2025-09-07T07:31:46.9431736Z d193fbc9ee81: Verifying Checksum 2025-09-07T07:31:46.9433048Z d193fbc9ee81: Download complete 2025-09-07T07:33:25.3204530Z d193fbc9ee81: Pull complete 2025-09-07T07:33:25.5499332Z 99a2462c8c29: Pull complete 2025-09-07T07:33:25.7926093Z fe1294e18ed0: Pull complete 2025-09-07T07:33:26.2586981Z f379f1cd3010: Pull complete 2025-09-07T07:33:26.4722892Z 6ae93e16ed4e: Pull complete 2025-09-07T07:33:27.2594061Z f995c3109bd8: Pull complete 2025-09-07T07:33:28.1916525Z 058384795bac: Pull complete 2025-09-07T07:33:28.5500783Z 3e762f827dd1: Pull complete 2025-09-07T07:33:29.4045359Z 6acfbf2cdc17: Pull complete 2025-09-07T07:33:30.4845740Z 1e43a43b4f06: Pull complete 2025-09-07T07:33:31.0060255Z 968c538d5104: Pull complete 2025-09-07T07:33:32.0565134Z 575807ea3b5f: Pull complete 2025-09-07T07:33:32.5290722Z 6a61bbb32004: Pull complete 2025-09-07T07:33:36.5634016Z f88dbb5c90de: Pull complete 2025-09-07T07:33:36.9335779Z 08443826e2c5: Pull complete 2025-09-07T07:33:37.4059275Z 923b78b30294: Pull complete 2025-09-07T07:33:42.2077299Z 0751aba37e5a: Pull complete 2025-09-07T07:33:42.6992159Z fbe315c5491d: Pull complete 2025-09-07T07:33:43.2345692Z af69784d4cfc: Pull complete 2025-09-07T07:33:47.9368537Z 91e4deab4f89: Pull complete 2025-09-07T07:33:48.4129829Z c1d56a01ea40: Pull complete 2025-09-07T07:33:48.7685117Z 9ad931d990d4: Pull complete 2025-09-07T07:33:49.1297645Z df5b3d079e9e: Pull complete 2025-09-07T07:33:49.5201561Z 24f5a79527e3: Pull complete 2025-09-07T07:33:49.7121128Z 4166cbf927c3: Pull complete 2025-09-07T07:33:50.7961690Z d243494d377d: Pull complete 2025-09-07T07:33:52.1640040Z 8a28267fa332: Pull complete 2025-09-07T07:33:52.6834559Z 89735be27bd4: Pull complete 2025-09-07T07:33:52.8867669Z Digest: sha256:c8bf3aa0f88e5662ff72435fb728fa2432ad59709b953fd68e95432215dfee4a 2025-09-07T07:33:52.9688440Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:33:53.0141761Z 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:33:53.0202068Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:33:53.0203029Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:33:53.0213379Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:33:53.0213696Z env: 2025-09-07T07:33:53.0213875Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:33:53.0214087Z ##[endgroup] 2025-09-07T07:33:53.0297663Z Prepare all required actions 2025-09-07T07:33:53.0347651Z ##[group]Run ./.github/actions/get-workflow-job-id 2025-09-07T07:33:53.0347938Z with: 2025-09-07T07:33:53.0348476Z github-token: *** 2025-09-07T07:33:53.0348676Z env: 2025-09-07T07:33:53.0348854Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:33:53.0349071Z ##[endgroup] 2025-09-07T07:33:53.0526716Z ##[group]Run set -eux 2025-09-07T07:33:53.0526938Z set -eux 2025-09-07T07:33:53.0527318Z python3 .github/scripts/get_workflow_job_id.py "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-09-07T07:33:53.0533241Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:33:53.0533562Z env: 2025-09-07T07:33:53.0533735Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:33:53.0534121Z GITHUB_TOKEN: *** 2025-09-07T07:33:53.0534312Z ##[endgroup] 2025-09-07T07:33:53.0556272Z + python3 .github/scripts/get_workflow_job_id.py 17525321311 i-0c7ee175fda0aa6a4 2025-09-07T07:33:54.1597162Z Setting output job-id=49775354812 2025-09-07T07:33:54.1597846Z Setting output job-name=linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal) 2025-09-07T07:33:54.1745417Z ##[group]Run python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-09-07T07:33:54.1746118Z python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-09-07T07:33:54.1746944Z python3 -m tools.stats.monitor --log-interval "$MONITOR_LOG_INTERVAL" --data-collect-interval "$MONITOR_DATA_COLLECT_INTERVAL" > usage_log.txt 2>&1 & 2025-09-07T07:33:54.1747677Z echo "monitor-script-pid=${!}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:33:54.1754417Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:33:54.1754730Z env: 2025-09-07T07:33:54.1754903Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:33:54.1755117Z JOB_ID: 49775354812 2025-09-07T07:33:54.1755607Z JOB_NAME: linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal) 2025-09-07T07:33:54.1756204Z WORKFLOW_NAME: inductor-perf-nightly-aarch64 2025-09-07T07:33:54.1756515Z WORKFLOW_RUN_ID: 17525321311 2025-09-07T07:33:54.1756746Z MONITOR_LOG_INTERVAL: 15 2025-09-07T07:33:54.1756973Z MONITOR_DATA_COLLECT_INTERVAL: 4 2025-09-07T07:33:54.1757205Z ##[endgroup] 2025-09-07T07:33:54.6993981Z Defaulting to user installation because normal site-packages is not writeable 2025-09-07T07:33:54.9928979Z Collecting psutil==5.9.8 2025-09-07T07:33:55.0130877Z Downloading psutil-5.9.8.tar.gz (503 kB) 2025-09-07T07:33:55.2563860Z Installing build dependencies: started 2025-09-07T07:33:57.7997187Z Installing build dependencies: finished with status 'done' 2025-09-07T07:33:57.8024118Z Getting requirements to build wheel: started 2025-09-07T07:33:58.1948548Z Getting requirements to build wheel: finished with status 'done' 2025-09-07T07:33:58.1958541Z Preparing metadata (pyproject.toml): started 2025-09-07T07:33:58.7507723Z Preparing metadata (pyproject.toml): finished with status 'done' 2025-09-07T07:33:58.8202447Z Collecting dataclasses_json==0.6.7 2025-09-07T07:33:58.8237533Z Downloading dataclasses_json-0.6.7-py3-none-any.whl (28 kB) 2025-09-07T07:33:58.9105102Z Collecting nvidia-ml-py==11.525.84 2025-09-07T07:33:58.9137912Z Downloading nvidia_ml_py-11.525.84-py3-none-any.whl (34 kB) 2025-09-07T07:33:58.9994383Z Collecting typing-inspect<1,>=0.4.0 2025-09-07T07:33:59.0027748Z Downloading typing_inspect-0.9.0-py3-none-any.whl (8.8 kB) 2025-09-07T07:33:59.1583459Z Collecting marshmallow<4.0.0,>=3.18.0 2025-09-07T07:33:59.1615922Z Downloading marshmallow-3.26.1-py3-none-any.whl (50 kB) 2025-09-07T07:33:59.2827641Z Collecting packaging>=17.0 2025-09-07T07:33:59.2859630Z Downloading packaging-25.0-py3-none-any.whl (66 kB) 2025-09-07T07:33:59.3569117Z Collecting mypy-extensions>=0.3.0 2025-09-07T07:33:59.3600407Z Downloading mypy_extensions-1.1.0-py3-none-any.whl (5.0 kB) 2025-09-07T07:33:59.4567441Z Collecting typing-extensions>=3.7.4 2025-09-07T07:33:59.4598727Z Downloading typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-09-07T07:33:59.4948681Z Building wheels for collected packages: psutil 2025-09-07T07:33:59.4954792Z Building wheel for psutil (pyproject.toml): started 2025-09-07T07:33:59.7273516Z Building wheel for psutil (pyproject.toml): finished with status 'error' 2025-09-07T07:33:59.7274631Z ERROR: Command errored out with exit status 1: 2025-09-07T07:33:59.7275333Z command: /usr/bin/python3 /usr/lib/python3.9/site-packages/pip/_vendor/pep517/in_process/_in_process.py build_wheel /tmp/tmpaks_hp7_ 2025-09-07T07:33:59.7276064Z cwd: /tmp/pip-install-2ibwilx0/psutil_f5840ea8862241df8d1553d806551ed0 2025-09-07T07:33:59.7276446Z Complete output (59 lines): 2025-09-07T07:33:59.7277145Z /tmp/pip-build-env-69keg7ms/overlay/lib/python3.9/site-packages/setuptools/dist.py:759: SetuptoolsDeprecationWarning: License classifiers are deprecated. 2025-09-07T07:33:59.7277854Z !! 2025-09-07T07:33:59.7278005Z 2025-09-07T07:33:59.7278198Z ******************************************************************************** 2025-09-07T07:33:59.7279086Z Please consider removing the following classifiers in favor of a SPDX license expression: 2025-09-07T07:33:59.7279585Z 2025-09-07T07:33:59.7279812Z License :: OSI Approved :: BSD License 2025-09-07T07:33:59.7280073Z 2025-09-07T07:33:59.7280494Z See https://packaging.python.org/en/latest/guides/writing-pyproject-toml/#license for details. 2025-09-07T07:33:59.7281001Z ******************************************************************************** 2025-09-07T07:33:59.7281267Z 2025-09-07T07:33:59.7281415Z !! 2025-09-07T07:33:59.7281602Z self._finalize_license_expression() 2025-09-07T07:33:59.7281864Z running bdist_wheel 2025-09-07T07:33:59.7282062Z running build 2025-09-07T07:33:59.7282250Z running build_py 2025-09-07T07:33:59.7282503Z creating build/lib.linux-aarch64-cpython-39/psutil 2025-09-07T07:33:59.7282953Z copying psutil/_pswindows.py -> build/lib.linux-aarch64-cpython-39/psutil 2025-09-07T07:33:59.7283493Z copying psutil/_pssunos.py -> build/lib.linux-aarch64-cpython-39/psutil 2025-09-07T07:33:59.7283999Z copying psutil/_psposix.py -> build/lib.linux-aarch64-cpython-39/psutil 2025-09-07T07:33:59.7284490Z copying psutil/_psosx.py -> build/lib.linux-aarch64-cpython-39/psutil 2025-09-07T07:33:59.7284985Z copying psutil/_pslinux.py -> build/lib.linux-aarch64-cpython-39/psutil 2025-09-07T07:33:59.7285476Z copying psutil/_psbsd.py -> build/lib.linux-aarch64-cpython-39/psutil 2025-09-07T07:33:59.7285958Z copying psutil/_psaix.py -> build/lib.linux-aarch64-cpython-39/psutil 2025-09-07T07:33:59.7286447Z copying psutil/_compat.py -> build/lib.linux-aarch64-cpython-39/psutil 2025-09-07T07:33:59.7286931Z copying psutil/_common.py -> build/lib.linux-aarch64-cpython-39/psutil 2025-09-07T07:33:59.7287420Z copying psutil/__init__.py -> build/lib.linux-aarch64-cpython-39/psutil 2025-09-07T07:33:59.7287877Z creating build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7288401Z copying psutil/tests/test_windows.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7289014Z copying psutil/tests/test_unicode.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7289638Z copying psutil/tests/test_testutils.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7290253Z copying psutil/tests/test_system.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7290855Z copying psutil/tests/test_sunos.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7291477Z copying psutil/tests/test_process_all.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7292098Z copying psutil/tests/test_process.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7292696Z copying psutil/tests/test_posix.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7293275Z copying psutil/tests/test_osx.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7294104Z copying psutil/tests/test_misc.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7294705Z copying psutil/tests/test_memleaks.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7295305Z copying psutil/tests/test_linux.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7295918Z copying psutil/tests/test_contracts.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7296554Z copying psutil/tests/test_connections.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7297161Z copying psutil/tests/test_bsd.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7297735Z copying psutil/tests/test_aix.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7298316Z copying psutil/tests/runner.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7298884Z copying psutil/tests/__main__.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7299600Z copying psutil/tests/__init__.py -> build/lib.linux-aarch64-cpython-39/psutil/tests 2025-09-07T07:33:59.7300013Z running build_ext 2025-09-07T07:33:59.7300237Z building 'psutil._psutil_linux' extension 2025-09-07T07:33:59.7300575Z creating build/temp.linux-aarch64-cpython-39/psutil 2025-09-07T07:33:59.7300984Z creating build/temp.linux-aarch64-cpython-39/psutil/arch/linux 2025-09-07T07:33:59.7305845Z gcc -Wno-unused-result -Wsign-compare -DDYNAMIC_ANNOTATIONS_ENABLED=1 -DNDEBUG -O2 -ftree-vectorize -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -march=armv8.2-a+crypto -mtune=neoverse-n1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -ftree-vectorize -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -march=armv8.2-a+crypto -mtune=neoverse-n1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -ftree-vectorize -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -march=armv8.2-a+crypto -mtune=neoverse-n1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -D_GNU_SOURCE -fPIC -fwrapv -fPIC -DPSUTIL_POSIX=1 -DPSUTIL_SIZEOF_PID_T=4 -DPSUTIL_VERSION=598 -DPy_LIMITED_API=0x03060000 -DPSUTIL_LINUX=1 -I/usr/include/python3.9 -c psutil/_psutil_common.c -o build/temp.linux-aarch64-cpython-39/psutil/_psutil_common.o 2025-09-07T07:33:59.7310825Z psutil/_psutil_common.c:9:10: fatal error: Python.h: No such file or directory 2025-09-07T07:33:59.7311225Z 9 | #include 2025-09-07T07:33:59.7311453Z | ^~~~~~~~~~ 2025-09-07T07:33:59.7311688Z compilation terminated. 2025-09-07T07:33:59.7312140Z psutil could not be installed from sources. Perhaps Python header files are not installed. Try running: 2025-09-07T07:33:59.7312632Z sudo yum install gcc python3-devel 2025-09-07T07:33:59.7312951Z error: command '/usr/bin/gcc' failed with exit code 1 2025-09-07T07:33:59.7313267Z ---------------------------------------- 2025-09-07T07:33:59.7313547Z ERROR: Failed building wheel for psutil 2025-09-07T07:33:59.7314018Z ERROR: Could not build wheels for psutil, which is required to install pyproject.toml-based projects 2025-09-07T07:33:59.7314477Z Failed to build psutil 2025-09-07T07:33:59.7963819Z ##[error]Process completed with exit code 1. 2025-09-07T07:33:59.8017695Z Prepare all required actions 2025-09-07T07:33:59.8018072Z Getting action download info 2025-09-07T07:33:59.9369176Z Download action repository 'seemethere/download-artifact-s3@v4' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-09-07T07:34:00.9781826Z Download action repository 'actions/download-artifact@v4' (SHA:d3f86a106a0bac45b974a628896c90dbdf5c8093) 2025-09-07T07:34:04.4750153Z ##[group]Run ./.github/actions/download-build-artifacts 2025-09-07T07:34:04.4750456Z with: 2025-09-07T07:34:04.4750648Z name: linux-jammy-aarch64-py3.10 2025-09-07T07:34:04.4750899Z s3-bucket: gha-artifacts 2025-09-07T07:34:04.4751107Z env: 2025-09-07T07:34:04.4751276Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:04.4751485Z ##[endgroup] 2025-09-07T07:34:04.4807641Z ##[group]Run seemethere/download-artifact-s3@v4 2025-09-07T07:34:04.4807926Z with: 2025-09-07T07:34:04.4808114Z name: linux-jammy-aarch64-py3.10 2025-09-07T07:34:04.4808375Z s3-bucket: gha-artifacts 2025-09-07T07:34:04.4808622Z region: us-east-1 2025-09-07T07:34:04.4808802Z env: 2025-09-07T07:34:04.4808968Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:04.4809179Z ##[endgroup] 2025-09-07T07:34:05.0973483Z (node:51933) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-09-07T07:34:05.0973945Z 2025-09-07T07:34:05.0974109Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-09-07T07:34:05.0974576Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-09-07T07:34:05.0975056Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-09-07T07:34:05.8381092Z Found 1 objects with prefix pytorch/pytorch/17525321311/linux-jammy-aarch64-py3.10/ 2025-09-07T07:34:05.8381720Z Starting download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-09-07T07:34:08.9035130Z Finished download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-09-07T07:34:08.9041179Z Artifact download has finished successfully 2025-09-07T07:34:08.9299411Z ##[group]Run unzip -o artifacts.zip 2025-09-07T07:34:08.9299702Z unzip -o artifacts.zip 2025-09-07T07:34:08.9306974Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:34:08.9307294Z env: 2025-09-07T07:34:08.9307470Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:08.9307704Z ##[endgroup] 2025-09-07T07:34:08.9611650Z Archive: artifacts.zip 2025-09-07T07:34:08.9612778Z creating: dist/ 2025-09-07T07:34:09.9019626Z inflating: dist/torch-2.9.0a0+git93fb23d-cp310-cp310-linux_aarch64.whl 2025-09-07T07:34:09.9020032Z creating: dist/vision/ 2025-09-07T07:34:09.9114563Z inflating: dist/vision/torchvision-0.22.0a0+966da7e-cp310-cp310-linux_aarch64.whl 2025-09-07T07:34:09.9115005Z creating: dist/audio/ 2025-09-07T07:34:09.9149546Z inflating: dist/audio/torchaudio-2.8.0a0+2e30055-cp310-cp310-linux_aarch64.whl 2025-09-07T07:34:09.9149958Z creating: dist/ao/ 2025-09-07T07:34:09.9194252Z inflating: dist/ao/torchao-0.7.0+git51c87b6e-py3-none-any.whl 2025-09-07T07:34:09.9320228Z inflating: dist/.ninja_log 2025-09-07T07:34:09.9320937Z creating: build/custom_test_artifacts/ 2025-09-07T07:34:09.9321328Z creating: build/custom_test_artifacts/custom-op-build/ 2025-09-07T07:34:09.9321757Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/ 2025-09-07T07:34:09.9322304Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/pkgRedirects/ 2025-09-07T07:34:09.9325328Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeConfigureLog.yaml 2025-09-07T07:34:09.9325909Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/ 2025-09-07T07:34:09.9326478Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeSystem.cmake 2025-09-07T07:34:09.9327089Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/ 2025-09-07T07:34:09.9327994Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/tmp/ 2025-09-07T07:34:09.9328847Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/CMakeCCompilerId.c 2025-09-07T07:34:09.9329963Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/a.out 2025-09-07T07:34:09.9330881Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeCCompiler.cmake 2025-09-07T07:34:09.9331522Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/ 2025-09-07T07:34:09.9332136Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/tmp/ 2025-09-07T07:34:09.9333398Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-09-07T07:34:09.9334442Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/a.out 2025-09-07T07:34:09.9335218Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeCXXCompiler.cmake 2025-09-07T07:34:09.9336334Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_C.bin 2025-09-07T07:34:09.9337699Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_CXX.bin 2025-09-07T07:34:09.9338381Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeScratch/ 2025-09-07T07:34:09.9338954Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/cmake.check_cache 2025-09-07T07:34:09.9339536Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/ 2025-09-07T07:34:09.9340172Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.ts 2025-09-07T07:34:09.9340899Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.make 2025-09-07T07:34:09.9341606Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/depend.make 2025-09-07T07:34:09.9342267Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/link.txt 2025-09-07T07:34:09.9342937Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/cmake_clean.cmake 2025-09-07T07:34:09.9343612Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/build.make 2025-09-07T07:34:09.9344295Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/DependInfo.cmake 2025-09-07T07:34:09.9344967Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/flags.make 2025-09-07T07:34:09.9345634Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/progress.make 2025-09-07T07:34:09.9364833Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o.d 2025-09-07T07:34:09.9599227Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o 2025-09-07T07:34:09.9599864Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/ 2025-09-07T07:34:09.9600548Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.ts 2025-09-07T07:34:09.9601318Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.make 2025-09-07T07:34:09.9602069Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/depend.make 2025-09-07T07:34:09.9602754Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/link.txt 2025-09-07T07:34:09.9603464Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/cmake_clean.cmake 2025-09-07T07:34:09.9604186Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/build.make 2025-09-07T07:34:09.9605118Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/DependInfo.cmake 2025-09-07T07:34:09.9605843Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/flags.make 2025-09-07T07:34:09.9606545Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/progress.make 2025-09-07T07:34:09.9626287Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o.d 2025-09-07T07:34:09.9719917Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o 2025-09-07T07:34:09.9720699Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-09-07T07:34:09.9721389Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/TargetDirectories.txt 2025-09-07T07:34:09.9722007Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/progress.marks 2025-09-07T07:34:09.9722585Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile2 2025-09-07T07:34:09.9723140Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile.cmake 2025-09-07T07:34:09.9723740Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/InstallScripts.json 2025-09-07T07:34:09.9724789Z inflating: build/custom_test_artifacts/custom-op-build/CMakeCache.txt 2025-09-07T07:34:09.9725586Z inflating: build/custom_test_artifacts/custom-op-build/Makefile 2025-09-07T07:34:09.9726085Z inflating: build/custom_test_artifacts/custom-op-build/cmake_install.cmake 2025-09-07T07:34:09.9910502Z inflating: build/custom_test_artifacts/custom-op-build/libcustom_ops.so 2025-09-07T07:34:09.9971780Z inflating: build/custom_test_artifacts/custom-op-build/test_custom_ops 2025-09-07T07:34:09.9972221Z creating: build/custom_test_artifacts/jit-hook-build/ 2025-09-07T07:34:09.9972632Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/ 2025-09-07T07:34:09.9973131Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/pkgRedirects/ 2025-09-07T07:34:09.9976267Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeConfigureLog.yaml 2025-09-07T07:34:09.9976835Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/ 2025-09-07T07:34:09.9977387Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeSystem.cmake 2025-09-07T07:34:09.9977998Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/ 2025-09-07T07:34:09.9978579Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/tmp/ 2025-09-07T07:34:09.9979682Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/CMakeCCompilerId.c 2025-09-07T07:34:09.9980776Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/a.out 2025-09-07T07:34:09.9981430Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeCCompiler.cmake 2025-09-07T07:34:09.9982060Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/ 2025-09-07T07:34:09.9982658Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/tmp/ 2025-09-07T07:34:09.9984414Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-09-07T07:34:09.9985395Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/a.out 2025-09-07T07:34:09.9986286Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeCXXCompiler.cmake 2025-09-07T07:34:09.9987414Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_C.bin 2025-09-07T07:34:09.9988693Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_CXX.bin 2025-09-07T07:34:09.9989356Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeScratch/ 2025-09-07T07:34:09.9990138Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/cmake.check_cache 2025-09-07T07:34:09.9990737Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/ 2025-09-07T07:34:09.9991394Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.ts 2025-09-07T07:34:09.9992292Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.make 2025-09-07T07:34:09.9993010Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/depend.make 2025-09-07T07:34:09.9993691Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/link.txt 2025-09-07T07:34:09.9994382Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/cmake_clean.cmake 2025-09-07T07:34:09.9995077Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/build.make 2025-09-07T07:34:09.9995777Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/DependInfo.cmake 2025-09-07T07:34:09.9996470Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/flags.make 2025-09-07T07:34:09.9997155Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/progress.make 2025-09-07T07:34:10.0015752Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o.d 2025-09-07T07:34:10.0088944Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o 2025-09-07T07:34:10.0089692Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-09-07T07:34:10.0090361Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/TargetDirectories.txt 2025-09-07T07:34:10.0090963Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/progress.marks 2025-09-07T07:34:10.0091523Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile2 2025-09-07T07:34:10.0092067Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile.cmake 2025-09-07T07:34:10.0092657Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/InstallScripts.json 2025-09-07T07:34:10.0093728Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeCache.txt 2025-09-07T07:34:10.0094446Z inflating: build/custom_test_artifacts/jit-hook-build/Makefile 2025-09-07T07:34:10.0094959Z inflating: build/custom_test_artifacts/jit-hook-build/cmake_install.cmake 2025-09-07T07:34:10.0137036Z inflating: build/custom_test_artifacts/jit-hook-build/test_jit_hooks 2025-09-07T07:34:10.0137494Z creating: build/custom_test_artifacts/custom-backend-build/ 2025-09-07T07:34:10.0137951Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/ 2025-09-07T07:34:10.0138503Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/pkgRedirects/ 2025-09-07T07:34:10.0141473Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeConfigureLog.yaml 2025-09-07T07:34:10.0142093Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/ 2025-09-07T07:34:10.0142695Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeSystem.cmake 2025-09-07T07:34:10.0143354Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/ 2025-09-07T07:34:10.0144010Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/tmp/ 2025-09-07T07:34:10.0144838Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/CMakeCCompilerId.c 2025-09-07T07:34:10.0145946Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/a.out 2025-09-07T07:34:10.0146822Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeCCompiler.cmake 2025-09-07T07:34:10.0147512Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/ 2025-09-07T07:34:10.0148167Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/tmp/ 2025-09-07T07:34:10.0149540Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-09-07T07:34:10.0150568Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/a.out 2025-09-07T07:34:10.0151415Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeCXXCompiler.cmake 2025-09-07T07:34:10.0152523Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_C.bin 2025-09-07T07:34:10.0153808Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_CXX.bin 2025-09-07T07:34:10.0154804Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeScratch/ 2025-09-07T07:34:10.0155479Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/cmake.check_cache 2025-09-07T07:34:10.0156162Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/ 2025-09-07T07:34:10.0157000Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.ts 2025-09-07T07:34:10.0176499Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.make 2025-09-07T07:34:10.0177430Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/depend.make 2025-09-07T07:34:10.0178179Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/link.txt 2025-09-07T07:34:10.0178941Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/cmake_clean.cmake 2025-09-07T07:34:10.0179732Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/build.make 2025-09-07T07:34:10.0180516Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/DependInfo.cmake 2025-09-07T07:34:10.0181273Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/flags.make 2025-09-07T07:34:10.0182037Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/progress.make 2025-09-07T07:34:10.0182843Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o.d 2025-09-07T07:34:10.0302778Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o 2025-09-07T07:34:10.0303532Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/ 2025-09-07T07:34:10.0304295Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.ts 2025-09-07T07:34:10.0305144Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.make 2025-09-07T07:34:10.0305968Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/depend.make 2025-09-07T07:34:10.0306739Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/link.txt 2025-09-07T07:34:10.0307521Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/cmake_clean.cmake 2025-09-07T07:34:10.0308315Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/build.make 2025-09-07T07:34:10.0309112Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/DependInfo.cmake 2025-09-07T07:34:10.0310152Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/flags.make 2025-09-07T07:34:10.0310946Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/progress.make 2025-09-07T07:34:10.0329499Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o.d 2025-09-07T07:34:10.0392209Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o 2025-09-07T07:34:10.0393050Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-09-07T07:34:10.0393775Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/TargetDirectories.txt 2025-09-07T07:34:10.0394430Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/progress.marks 2025-09-07T07:34:10.0395038Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile2 2025-09-07T07:34:10.0395641Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile.cmake 2025-09-07T07:34:10.0396281Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/InstallScripts.json 2025-09-07T07:34:10.0397248Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeCache.txt 2025-09-07T07:34:10.0397869Z inflating: build/custom_test_artifacts/custom-backend-build/Makefile 2025-09-07T07:34:10.0398491Z inflating: build/custom_test_artifacts/custom-backend-build/cmake_install.cmake 2025-09-07T07:34:10.0505972Z inflating: build/custom_test_artifacts/custom-backend-build/libcustom_backend.so 2025-09-07T07:34:10.0548785Z inflating: build/custom_test_artifacts/custom-backend-build/test_custom_backend 2025-09-07T07:34:10.0549198Z creating: build/lib/ 2025-09-07T07:34:10.0635829Z inflating: build/lib/libprotobuf-lite.a 2025-09-07T07:34:10.1089269Z inflating: build/lib/libprotobuf.a 2025-09-07T07:34:10.1095713Z inflating: build/lib/libpthreadpool.a 2025-09-07T07:34:10.1569373Z inflating: build/lib/libprotoc.a 2025-09-07T07:34:10.1578464Z inflating: build/lib/libcpuinfo.a 2025-09-07T07:34:10.1586484Z inflating: build/lib/libcpuinfo_internals.a 2025-09-07T07:34:10.1587382Z inflating: build/lib/libclog.a 2025-09-07T07:34:10.1606803Z inflating: build/lib/libpytorch_qnnpack.a 2025-09-07T07:34:10.1624440Z inflating: build/lib/libnnpack.a 2025-09-07T07:34:10.1626319Z inflating: build/lib/libnnpack_reference_layers.a 2025-09-07T07:34:10.1745976Z inflating: build/lib/libmicrokernels-prod.a 2025-09-07T07:34:10.2220778Z inflating: build/lib/libmicrokernels-all.a 2025-09-07T07:34:10.2291944Z inflating: build/lib/libgtest.a 2025-09-07T07:34:10.2309635Z inflating: build/lib/libgmock.a 2025-09-07T07:34:10.2310329Z inflating: build/lib/libgtest_main.a 2025-09-07T07:34:10.2311031Z inflating: build/lib/libgmock_main.a 2025-09-07T07:34:10.2396073Z inflating: build/lib/libXNNPACK.a 2025-09-07T07:34:10.2470037Z inflating: build/lib/libbenchmark.a 2025-09-07T07:34:10.2470760Z inflating: build/lib/libbenchmark_main.a 2025-09-07T07:34:10.2501592Z inflating: build/lib/libtensorpipe_uv.a 2025-09-07T07:34:10.3112844Z inflating: build/lib/libtensorpipe.a 2025-09-07T07:34:10.3236799Z inflating: build/lib/libgloo.a 2025-09-07T07:34:10.3285767Z inflating: build/lib/libonnx_proto.a 2025-09-07T07:34:10.4019850Z inflating: build/lib/libonnx.a 2025-09-07T07:34:11.0864593Z inflating: build/lib/libdnnl.a 2025-09-07T07:34:11.0930173Z inflating: build/lib/libkleidiai.a 2025-09-07T07:34:11.0948453Z inflating: build/lib/libfmt.a 2025-09-07T07:34:11.1210652Z inflating: build/lib/libkineto.a 2025-09-07T07:34:11.1319707Z inflating: build/lib/libc10.so 2025-09-07T07:34:11.1320777Z inflating: build/lib/libtorch_global_deps.so 2025-09-07T07:34:11.1326487Z inflating: build/lib/libCaffe2_perfkernels_sve.a 2025-09-07T07:34:12.8189591Z inflating: build/lib/libtorch_cpu.so 2025-09-07T07:34:12.8191120Z inflating: build/lib/libtorch.so 2025-09-07T07:34:12.8254558Z inflating: build/lib/libtorchbind_test.so 2025-09-07T07:34:12.8273423Z inflating: build/lib/libjitbackend_test.so 2025-09-07T07:34:12.8296525Z inflating: build/lib/libbackend_with_compiler.so 2025-09-07T07:34:12.8323591Z inflating: build/lib/libaoti_custom_ops.so 2025-09-07T07:34:12.8327032Z inflating: build/lib/libshm.so 2025-09-07T07:34:13.0285650Z inflating: build/lib/libtorch_python.so 2025-09-07T07:34:13.0319934Z inflating: build/lib/libnnapi_backend.so 2025-09-07T07:34:13.0320211Z creating: build/bin/ 2025-09-07T07:34:13.0320437Z creating: build/bin/CMakeFiles/ 2025-09-07T07:34:13.0321028Z inflating: build/bin/cmake_install.cmake 2025-09-07T07:34:13.0321352Z inflating: build/bin/CTestTestfile.cmake 2025-09-07T07:34:13.0730815Z inflating: build/bin/protoc-3.13.0.0 2025-09-07T07:34:13.1141273Z inflating: build/bin/protoc 2025-09-07T07:34:13.1196203Z inflating: build/bin/c10_AllocatorConfig_test 2025-09-07T07:34:13.1247760Z inflating: build/bin/c10_CompileTimeFunctionPointer_test 2025-09-07T07:34:13.1301212Z inflating: build/bin/c10_DeviceGuard_test 2025-09-07T07:34:13.1354325Z inflating: build/bin/c10_Device_test 2025-09-07T07:34:13.1414875Z inflating: build/bin/c10_DispatchKeySet_test 2025-09-07T07:34:13.1465092Z inflating: build/bin/c10_StreamGuard_test 2025-09-07T07:34:13.1520491Z inflating: build/bin/c10_Scalar_test 2025-09-07T07:34:13.1579139Z inflating: build/bin/c10_SymInt_test 2025-09-07T07:34:13.1634760Z inflating: build/bin/c10_InlineDeviceGuard_test 2025-09-07T07:34:13.1691993Z inflating: build/bin/c10_InlineStreamGuard_test 2025-09-07T07:34:13.1749747Z inflating: build/bin/c10_SizesAndStrides_test 2025-09-07T07:34:13.1800156Z inflating: build/bin/c10_ConstexprCrc_test 2025-09-07T07:34:13.1854515Z inflating: build/bin/c10_Bitset_test 2025-09-07T07:34:13.1905395Z inflating: build/bin/c10_ArrayRef_test 2025-09-07T07:34:13.1956710Z inflating: build/bin/c10_DeadlockDetection_test 2025-09-07T07:34:13.2028366Z inflating: build/bin/c10_cow_test 2025-09-07T07:34:13.2087108Z inflating: build/bin/c10_Enumerate_test 2025-09-07T07:34:13.2138920Z inflating: build/bin/c10_Half_test 2025-09-07T07:34:13.2192834Z inflating: build/bin/c10_IntrusiveList_test 2025-09-07T07:34:13.2250125Z inflating: build/bin/c10_LeftRight_test 2025-09-07T07:34:13.2304253Z inflating: build/bin/c10_NetworkFlow_test 2025-09-07T07:34:13.2360466Z inflating: build/bin/c10_Metaprogramming_test 2025-09-07T07:34:13.2411884Z inflating: build/bin/c10_Semaphore_test 2025-09-07T07:34:13.2464889Z inflating: build/bin/c10_TypeIndex_test 2025-09-07T07:34:13.2521850Z inflating: build/bin/c10_ThreadLocal_test 2025-09-07T07:34:13.2574407Z inflating: build/bin/c10_TypeList_test 2025-09-07T07:34:13.2627195Z inflating: build/bin/c10_accumulate_test 2025-09-07T07:34:13.2678743Z inflating: build/bin/c10_Synchronized_test 2025-09-07T07:34:13.2729028Z inflating: build/bin/c10_TypeTraits_test 2025-09-07T07:34:13.2780561Z inflating: build/bin/c10_bit_cast_test 2025-09-07T07:34:13.2837326Z inflating: build/bin/c10_bfloat16_test 2025-09-07T07:34:13.2896121Z inflating: build/bin/c10_complex_math_test 2025-09-07T07:34:13.2947044Z inflating: build/bin/c10_error_test 2025-09-07T07:34:13.3003137Z inflating: build/bin/c10_complex_test 2025-09-07T07:34:13.3054859Z inflating: build/bin/c10_flags_test 2025-09-07T07:34:13.3109268Z inflating: build/bin/c10_exception_test 2025-09-07T07:34:13.3161992Z inflating: build/bin/c10_irange_test 2025-09-07T07:34:13.3217181Z inflating: build/bin/c10_lazy_test 2025-09-07T07:34:13.3269274Z inflating: build/bin/c10_generic_math_test 2025-09-07T07:34:13.3327468Z inflating: build/bin/c10_logging_test 2025-09-07T07:34:13.3381995Z inflating: build/bin/c10_registry_test 2025-09-07T07:34:13.3457499Z inflating: build/bin/c10_optional_test 2025-09-07T07:34:13.3520435Z inflating: build/bin/c10_ordered_preserving_dict_test 2025-09-07T07:34:13.3693884Z inflating: build/bin/c10_intrusive_ptr_test 2025-09-07T07:34:13.3747402Z inflating: build/bin/c10_ssize_test 2025-09-07T07:34:13.3896307Z inflating: build/bin/c10_small_vector_test 2025-09-07T07:34:13.3953269Z inflating: build/bin/c10_string_util_test 2025-09-07T07:34:13.3997054Z inflating: build/bin/c10_intrusive_ptr_benchmark 2025-09-07T07:34:13.4048735Z inflating: build/bin/c10_tempfile_test 2025-09-07T07:34:13.4099198Z inflating: build/bin/c10_string_view_test 2025-09-07T07:34:13.4156073Z inflating: build/bin/c10_typeid_test 2025-09-07T07:34:13.4460628Z inflating: build/bin/vec_test_all_types_DEFAULT 2025-09-07T07:34:13.4754394Z inflating: build/bin/vec_test_all_types_SVE256 2025-09-07T07:34:13.4808208Z inflating: build/bin/static_runtime_bench 2025-09-07T07:34:13.5064729Z inflating: build/bin/static_runtime_test 2025-09-07T07:34:13.5141170Z inflating: build/bin/Dict_test 2025-09-07T07:34:13.5194602Z inflating: build/bin/Dimname_test 2025-09-07T07:34:13.5260546Z inflating: build/bin/MaybeOwned_test 2025-09-07T07:34:13.5318479Z inflating: build/bin/NamedTensor_test 2025-09-07T07:34:13.5378119Z inflating: build/bin/apply_utils_test 2025-09-07T07:34:13.5438561Z inflating: build/bin/atest 2025-09-07T07:34:13.5503051Z inflating: build/bin/basic 2025-09-07T07:34:13.5559459Z inflating: build/bin/broadcast_test 2025-09-07T07:34:13.5611582Z inflating: build/bin/cpu_allocator_test 2025-09-07T07:34:13.5670983Z inflating: build/bin/cpu_generator_test 2025-09-07T07:34:13.5725851Z inflating: build/bin/cpu_profiling_allocator_test 2025-09-07T07:34:13.5815355Z inflating: build/bin/cpu_rng_test 2025-09-07T07:34:13.5868315Z inflating: build/bin/dlconvertor_test 2025-09-07T07:34:13.5927675Z inflating: build/bin/extension_backend_test 2025-09-07T07:34:13.5984244Z inflating: build/bin/half_test 2025-09-07T07:34:13.6081562Z inflating: build/bin/ivalue_test 2025-09-07T07:34:13.6132959Z inflating: build/bin/lazy_tensor_test 2025-09-07T07:34:13.6188049Z inflating: build/bin/math_kernel_test 2025-09-07T07:34:13.6243049Z inflating: build/bin/memory_format_test 2025-09-07T07:34:13.6297952Z inflating: build/bin/memory_overlapping_test 2025-09-07T07:34:13.6352271Z inflating: build/bin/mobile_memory_cleanup 2025-09-07T07:34:13.6409253Z inflating: build/bin/native_test 2025-09-07T07:34:13.6461010Z inflating: build/bin/operator_name_test 2025-09-07T07:34:13.6513379Z inflating: build/bin/operators_test 2025-09-07T07:34:13.6566678Z inflating: build/bin/packedtensoraccessor_test 2025-09-07T07:34:13.6633569Z inflating: build/bin/pow_test 2025-09-07T07:34:13.6693092Z inflating: build/bin/quantized_test 2025-09-07T07:34:13.6744099Z inflating: build/bin/reduce_ops_test 2025-09-07T07:34:13.6796621Z inflating: build/bin/reportMemoryUsage_test 2025-09-07T07:34:13.6854223Z inflating: build/bin/scalar_tensor_test 2025-09-07T07:34:13.6913091Z inflating: build/bin/scalar_test 2025-09-07T07:34:13.6965923Z inflating: build/bin/StorageUtils_test 2025-09-07T07:34:13.7019259Z inflating: build/bin/stride_properties_test 2025-09-07T07:34:13.7100436Z inflating: build/bin/tensor_iterator_test 2025-09-07T07:34:13.7156571Z inflating: build/bin/test_parallel 2025-09-07T07:34:13.7208656Z inflating: build/bin/thread_init_test 2025-09-07T07:34:13.7264286Z inflating: build/bin/type_ptr_test 2025-09-07T07:34:13.7324930Z inflating: build/bin/type_test 2025-09-07T07:34:13.7378564Z inflating: build/bin/undefined_tensor_test 2025-09-07T07:34:13.7429430Z inflating: build/bin/verify_api_visibility 2025-09-07T07:34:13.7498860Z inflating: build/bin/legacy_vmap_test 2025-09-07T07:34:13.7551291Z inflating: build/bin/weakref_test 2025-09-07T07:34:13.7603947Z inflating: build/bin/wrapdim_test 2025-09-07T07:34:13.7657017Z inflating: build/bin/xla_tensor_test 2025-09-07T07:34:13.7718115Z inflating: build/bin/IListRef_test 2025-09-07T07:34:13.7829381Z inflating: build/bin/List_test 2025-09-07T07:34:13.7896892Z inflating: build/bin/KernelFunction_test 2025-09-07T07:34:13.8017378Z inflating: build/bin/kernel_function_legacy_test 2025-09-07T07:34:13.8112721Z inflating: build/bin/kernel_function_test 2025-09-07T07:34:13.8214448Z inflating: build/bin/kernel_lambda_test 2025-09-07T07:34:13.8339288Z inflating: build/bin/kernel_lambda_legacy_test 2025-09-07T07:34:13.8400991Z inflating: build/bin/kernel_stackbased_test 2025-09-07T07:34:13.8495976Z inflating: build/bin/make_boxed_from_unboxed_functor_test 2025-09-07T07:34:13.8548453Z inflating: build/bin/CppSignature_test 2025-09-07T07:34:13.8604863Z inflating: build/bin/backend_fallback_test 2025-09-07T07:34:13.8654961Z inflating: build/bin/op_allowlist_test 2025-09-07T07:34:13.8947588Z inflating: build/bin/op_registration_test 2025-09-07T07:34:13.9014598Z inflating: build/bin/inline_container_test 2025-09-07T07:34:14.0065472Z inflating: build/bin/test_jit 2025-09-07T07:34:14.0119770Z inflating: build/bin/FileStoreTest 2025-09-07T07:34:14.0173606Z inflating: build/bin/BackoffTest 2025-09-07T07:34:14.0231786Z inflating: build/bin/TCPStoreTest 2025-09-07T07:34:14.0286729Z inflating: build/bin/HashStoreTest 2025-09-07T07:34:14.0649147Z inflating: build/bin/test_nativert 2025-09-07T07:34:14.0652461Z inflating: build/bin/example_allreduce 2025-09-07T07:34:14.0720050Z inflating: build/bin/ProcessGroupGlooTest 2025-09-07T07:34:14.0776438Z inflating: build/bin/test_dist_autograd 2025-09-07T07:34:14.0845613Z inflating: build/bin/test_cpp_rpc 2025-09-07T07:34:14.1927106Z inflating: build/bin/test_api 2025-09-07T07:34:14.1929359Z inflating: build/bin/parallel_benchmark 2025-09-07T07:34:14.2279505Z inflating: build/bin/test_lazy 2025-09-07T07:34:14.2283285Z inflating: build/bin/torch_shm_manager 2025-09-07T07:34:14.2283574Z creating: .additional_ci_files/ 2025-09-07T07:34:14.2377273Z inflating: .additional_ci_files/test-times.json 2025-09-07T07:34:14.2730224Z inflating: .additional_ci_files/test-class-times.json 2025-09-07T07:34:14.2786433Z ##[group]Run rm artifacts.zip 2025-09-07T07:34:14.2786709Z rm artifacts.zip 2025-09-07T07:34:14.2793308Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:34:14.2793623Z env: 2025-09-07T07:34:14.2793798Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:14.2794010Z ##[endgroup] 2025-09-07T07:34:14.3453350Z ##[group]Run df -H 2025-09-07T07:34:14.3453552Z df -H 2025-09-07T07:34:14.3459211Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:34:14.3459527Z env: 2025-09-07T07:34:14.3459705Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:14.3459916Z ##[endgroup] 2025-09-07T07:34:14.3499943Z Filesystem Size Used Avail Use% Mounted on 2025-09-07T07:34:14.3500274Z devtmpfs 4.2M 0 4.2M 0% /dev 2025-09-07T07:34:14.3500552Z tmpfs 135G 0 135G 0% /dev/shm 2025-09-07T07:34:14.3500834Z tmpfs 54G 9.4M 54G 1% /run 2025-09-07T07:34:14.3501103Z /dev/nvme0n1p1 275G 46G 229G 17% / 2025-09-07T07:34:14.3501371Z tmpfs 135G 13k 135G 1% /tmp 2025-09-07T07:34:14.3501668Z /dev/nvme0n1p128 11M 1.5M 9.1M 14% /boot/efi 2025-09-07T07:34:14.3532933Z Prepare all required actions 2025-09-07T07:34:14.3533830Z Getting action download info 2025-09-07T07:34:14.4713473Z ##[group]Run ./.github/actions/download-td-artifacts 2025-09-07T07:34:14.4713772Z with: 2025-09-07T07:34:14.4713935Z env: 2025-09-07T07:34:14.4714107Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:14.4714324Z ##[endgroup] 2025-09-07T07:34:14.4931151Z ##[group]Run seemethere/download-artifact-s3@v4 2025-09-07T07:34:14.4931434Z with: 2025-09-07T07:34:14.4931599Z name: td_results 2025-09-07T07:34:14.4931793Z s3-bucket: gha-artifacts 2025-09-07T07:34:14.4932009Z region: us-east-1 2025-09-07T07:34:14.4932183Z env: 2025-09-07T07:34:14.4932354Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:14.4932563Z ##[endgroup] 2025-09-07T07:34:14.8411457Z (node:51952) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-09-07T07:34:14.8411886Z 2025-09-07T07:34:14.8412426Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-09-07T07:34:14.8412907Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-09-07T07:34:14.8413380Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-09-07T07:34:14.9297801Z Found 0 objects with prefix pytorch/pytorch/17525321311/td_results/ 2025-09-07T07:34:14.9304878Z Artifact download has finished successfully 2025-09-07T07:34:14.9520268Z ##[group]Run mkdir -p .additional_ci_files 2025-09-07T07:34:14.9520584Z mkdir -p .additional_ci_files 2025-09-07T07:34:14.9520956Z mv td_results.json .additional_ci_files/td_results.json || true 2025-09-07T07:34:14.9527449Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:34:14.9527760Z env: 2025-09-07T07:34:14.9527937Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:14.9528150Z ##[endgroup] 2025-09-07T07:34:14.9573547Z mv: cannot stat 'td_results.json': No such file or directory 2025-09-07T07:34:14.9610679Z ##[group]Run .github/scripts/parse_ref.py 2025-09-07T07:34:14.9611068Z .github/scripts/parse_ref.py 2025-09-07T07:34:14.9616773Z shell: /usr/bin/bash -e {0} 2025-09-07T07:34:14.9617004Z env: 2025-09-07T07:34:14.9617176Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:14.9617389Z ##[endgroup] 2025-09-07T07:34:15.0706432Z Setting output branch=main 2025-09-07T07:34:15.0841358Z Prepare all required actions 2025-09-07T07:34:15.0841737Z Getting action download info 2025-09-07T07:34:15.1894951Z ##[group]Run ./.github/actions/filter-test-configs 2025-09-07T07:34:15.1895248Z with: 2025-09-07T07:34:15.1895680Z github-token: *** 2025-09-07T07:34:15.1908317Z test-matrix: {"include": [{"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 1, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 2, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 3, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 4, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 5, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 6, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 7, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 8, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 9, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 1, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 2, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 3, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 4, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 5, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 6, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 7, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 8, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 9, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 10, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 11, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 12, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 13, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 14, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 15, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 1, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 2, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 3, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 4, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 5, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 6, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 7, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 8, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 9, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 10, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 11, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 12, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 13, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 14, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 15, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}]} 2025-09-07T07:34:15.1921377Z job-name: linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal) 2025-09-07T07:34:15.1921915Z env: 2025-09-07T07:34:15.1922087Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:15.1922299Z ##[endgroup] 2025-09-07T07:34:15.2041779Z ##[group]Run nick-fields/retry@v3.0.0 2025-09-07T07:34:15.2042032Z with: 2025-09-07T07:34:15.2042190Z shell: bash 2025-09-07T07:34:15.2042372Z timeout_minutes: 10 2025-09-07T07:34:15.2042569Z max_attempts: 5 2025-09-07T07:34:15.2042763Z retry_wait_seconds: 30 2025-09-07T07:34:15.2043441Z command: set -eux # PyYAML 6.0 doesn't work with MacOS x86 anymore # This must run on Python-3.7 (AmazonLinux2) so can't use request=3.32.2 python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-09-07T07:34:15.2044164Z polling_interval_seconds: 1 2025-09-07T07:34:15.2044398Z warning_on_retry: true 2025-09-07T07:34:15.2044617Z continue_on_error: false 2025-09-07T07:34:15.2044830Z env: 2025-09-07T07:34:15.2044996Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:15.2045369Z GITHUB_TOKEN: *** 2025-09-07T07:34:15.2045558Z ##[endgroup] 2025-09-07T07:34:15.3711755Z + python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-09-07T07:34:15.5826455Z Defaulting to user installation because normal site-packages is not writeable 2025-09-07T07:34:16.4074483Z Collecting requests==2.27.1 2025-09-07T07:34:16.4216984Z Downloading requests-2.27.1-py2.py3-none-any.whl (63 kB) 2025-09-07T07:34:16.6477619Z Collecting pyyaml==6.0.2 2025-09-07T07:34:16.6513224Z Downloading PyYAML-6.0.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (720 kB) 2025-09-07T07:34:16.7756035Z Requirement already satisfied: idna<4,>=2.5 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (2.10) 2025-09-07T07:34:16.9046408Z Collecting certifi>=2017.4.17 2025-09-07T07:34:16.9080095Z Downloading certifi-2025.8.3-py3-none-any.whl (161 kB) 2025-09-07T07:34:17.3581653Z Collecting charset-normalizer~=2.0.0 2025-09-07T07:34:17.3617408Z Downloading charset_normalizer-2.0.12-py3-none-any.whl (39 kB) 2025-09-07T07:34:17.4782105Z Requirement already satisfied: urllib3<1.27,>=1.21.1 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (1.25.10) 2025-09-07T07:34:17.5382982Z Installing collected packages: charset-normalizer, certifi, requests, pyyaml 2025-09-07T07:34:18.0124431Z Successfully installed certifi-2025.8.3 charset-normalizer-2.0.12 pyyaml-6.0.2 requests-2.27.1 2025-09-07T07:34:18.2738831Z Command completed after 1 attempt(s). 2025-09-07T07:34:18.2804764Z ##[group]Run set -x 2025-09-07T07:34:18.2804983Z set -x 2025-09-07T07:34:18.2805166Z  2025-09-07T07:34:18.2805486Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-09-07T07:34:18.2805896Z # in runner workspace 2025-09-07T07:34:18.2806222Z python3 "${GITHUB_ACTION_PATH}/../../scripts/parse_ref.py" 2025-09-07T07:34:18.2813584Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:34:18.2813900Z env: 2025-09-07T07:34:18.2814072Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:18.2814282Z ##[endgroup] 2025-09-07T07:34:18.2836829Z + python3 /home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/filter-test-configs/../../scripts/parse_ref.py 2025-09-07T07:34:18.3002680Z Setting output branch=main 2025-09-07T07:34:18.3061718Z ##[group]Run echo "Workflow: ${GITHUB_WORKFLOW}" 2025-09-07T07:34:18.3062085Z echo "Workflow: ${GITHUB_WORKFLOW}" 2025-09-07T07:34:18.3062369Z echo "Job name: ${JOB_NAME}" 2025-09-07T07:34:18.3062608Z  2025-09-07T07:34:18.3062934Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-09-07T07:34:18.3063343Z # in runner workspace 2025-09-07T07:34:18.3063708Z python3 "${GITHUB_ACTION_PATH}/../../scripts/filter_test_configs.py" \ 2025-09-07T07:34:18.3064119Z  --workflow "${GITHUB_WORKFLOW}" \ 2025-09-07T07:34:18.3064416Z  --job-name "${JOB_NAME}" \ 2025-09-07T07:34:18.3077271Z  --test-matrix "{"include": [{"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 1, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 2, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 3, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 4, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 5, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 6, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 7, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 8, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 9, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 1, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 2, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 3, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 4, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 5, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 6, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 7, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 8, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 9, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 10, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 11, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 12, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 13, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 14, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 15, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 1, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 2, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 3, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 4, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 5, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 6, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 7, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 8, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 9, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 10, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 11, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 12, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 13, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 14, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 15, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}]}" \ 2025-09-07T07:34:18.3089999Z  --selected-test-configs "" \ 2025-09-07T07:34:18.3090281Z  --pr-number "${PR_NUMBER}" \ 2025-09-07T07:34:18.3090543Z  --tag "${TAG}" \ 2025-09-07T07:34:18.3090786Z  --event-name "${EVENT_NAME}" \ 2025-09-07T07:34:18.3091061Z  --schedule "${SCHEDULE}" \ 2025-09-07T07:34:18.3091322Z  --branch "${HEAD_BRANCH}" 2025-09-07T07:34:18.3097136Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:34:18.3097462Z env: 2025-09-07T07:34:18.3097634Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:18.3098185Z GITHUB_TOKEN: *** 2025-09-07T07:34:18.3098691Z JOB_NAME: linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal) 2025-09-07T07:34:18.3099227Z PR_NUMBER: 2025-09-07T07:34:18.3099404Z TAG: 2025-09-07T07:34:18.3099573Z EVENT_NAME: schedule 2025-09-07T07:34:18.3099773Z SCHEDULE: 0 7 * * * 2025-09-07T07:34:18.3099973Z HEAD_BRANCH: main 2025-09-07T07:34:18.3100158Z ##[endgroup] 2025-09-07T07:34:18.3121467Z Workflow: inductor-perf-nightly-aarch64 2025-09-07T07:34:18.3122054Z Job name: linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal) 2025-09-07T07:34:18.4890012Z Setting output keep-going=True 2025-09-07T07:34:18.4890303Z Setting output ci-verbose-test-logs=False 2025-09-07T07:34:18.4890606Z Setting output ci-test-showlocals=False 2025-09-07T07:34:18.4891140Z Setting output ci-no-test-timeout=False 2025-09-07T07:34:18.4891417Z Setting output ci-no-td=False 2025-09-07T07:34:18.4891681Z Setting output ci-td-distributed=False 2025-09-07T07:34:18.4891962Z Setting output is-unstable=False 2025-09-07T07:34:18.4892228Z Setting output reenabled-issues= 2025-09-07T07:34:18.4905137Z Setting output test-matrix={"include": [{"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 1, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 2, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 3, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 4, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 5, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 6, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 7, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 8, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 9, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 1, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 2, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 3, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 4, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 5, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 6, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 7, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 8, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 9, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 10, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 11, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 12, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 13, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 14, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 15, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 1, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 2, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 3, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 4, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 5, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 6, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 7, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 8, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 9, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 10, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 11, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 12, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 13, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 14, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 15, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}]} 2025-09-07T07:34:18.4917878Z Setting output is-test-matrix-empty=False 2025-09-07T07:34:18.5026730Z ##[group]Run echo "Filtered matrix:" 2025-09-07T07:34:18.5027128Z echo "Filtered matrix:" 2025-09-07T07:34:18.5039742Z echo "{"include": [{"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 1, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 2, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 3, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 4, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 5, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 6, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 7, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 8, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_huggingface_perf_cpu_aarch64", "shard": 9, "num_shards": 9, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 1, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 2, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 3, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 4, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 5, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 6, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 7, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 8, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 9, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 10, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 11, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 12, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 13, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 14, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_timm_perf_cpu_aarch64", "shard": 15, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 1, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 2, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 3, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 4, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 5, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 6, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 7, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 8, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 9, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 10, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 11, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 12, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 13, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 14, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}, {"config": "inductor_torchbench_perf_cpu_aarch64", "shard": 15, "num_shards": 15, "runner": "linux.arm64.m7g.metal"}]}" 2025-09-07T07:34:18.5052391Z  2025-09-07T07:34:18.5052560Z echo 2025-09-07T07:34:18.5052790Z echo "Is the current job unstable? False" 2025-09-07T07:34:18.5053068Z  2025-09-07T07:34:18.5053233Z echo 2025-09-07T07:34:18.5053447Z echo "Is keep-going label set? True" 2025-09-07T07:34:18.5053714Z  2025-09-07T07:34:18.5053874Z echo 2025-09-07T07:34:18.5054067Z echo "Reenabled issues? " 2025-09-07T07:34:18.5060096Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:34:18.5060411Z env: 2025-09-07T07:34:18.5060597Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:18.5060811Z ##[endgroup] 2025-09-07T07:34:18.5082647Z Filtered matrix: 2025-09-07T07:34:18.5098008Z {include: [{config: inductor_huggingface_perf_cpu_aarch64, shard: 1, num_shards: 9, runner: linux.arm64.m7g.metal}, {config: inductor_huggingface_perf_cpu_aarch64, shard: 2, num_shards: 9, runner: linux.arm64.m7g.metal}, {config: inductor_huggingface_perf_cpu_aarch64, shard: 3, num_shards: 9, runner: linux.arm64.m7g.metal}, {config: inductor_huggingface_perf_cpu_aarch64, shard: 4, num_shards: 9, runner: linux.arm64.m7g.metal}, {config: inductor_huggingface_perf_cpu_aarch64, shard: 5, num_shards: 9, runner: linux.arm64.m7g.metal}, {config: inductor_huggingface_perf_cpu_aarch64, shard: 6, num_shards: 9, runner: linux.arm64.m7g.metal}, {config: inductor_huggingface_perf_cpu_aarch64, shard: 7, num_shards: 9, runner: linux.arm64.m7g.metal}, {config: inductor_huggingface_perf_cpu_aarch64, shard: 8, num_shards: 9, runner: linux.arm64.m7g.metal}, {config: inductor_huggingface_perf_cpu_aarch64, shard: 9, num_shards: 9, runner: linux.arm64.m7g.metal}, {config: inductor_timm_perf_cpu_aarch64, shard: 1, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_timm_perf_cpu_aarch64, shard: 2, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_timm_perf_cpu_aarch64, shard: 3, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_timm_perf_cpu_aarch64, shard: 4, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_timm_perf_cpu_aarch64, shard: 5, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_timm_perf_cpu_aarch64, shard: 6, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_timm_perf_cpu_aarch64, shard: 7, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_timm_perf_cpu_aarch64, shard: 8, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_timm_perf_cpu_aarch64, shard: 9, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_timm_perf_cpu_aarch64, shard: 10, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_timm_perf_cpu_aarch64, shard: 11, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_timm_perf_cpu_aarch64, shard: 12, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_timm_perf_cpu_aarch64, shard: 13, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_timm_perf_cpu_aarch64, shard: 14, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_timm_perf_cpu_aarch64, shard: 15, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_torchbench_perf_cpu_aarch64, shard: 1, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_torchbench_perf_cpu_aarch64, shard: 2, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_torchbench_perf_cpu_aarch64, shard: 3, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_torchbench_perf_cpu_aarch64, shard: 4, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_torchbench_perf_cpu_aarch64, shard: 5, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_torchbench_perf_cpu_aarch64, shard: 6, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_torchbench_perf_cpu_aarch64, shard: 7, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_torchbench_perf_cpu_aarch64, shard: 8, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_torchbench_perf_cpu_aarch64, shard: 9, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_torchbench_perf_cpu_aarch64, shard: 10, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_torchbench_perf_cpu_aarch64, shard: 11, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_torchbench_perf_cpu_aarch64, shard: 12, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_torchbench_perf_cpu_aarch64, shard: 13, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_torchbench_perf_cpu_aarch64, shard: 14, num_shards: 15, runner: linux.arm64.m7g.metal}, {config: inductor_torchbench_perf_cpu_aarch64, shard: 15, num_shards: 15, runner: linux.arm64.m7g.metal}]} 2025-09-07T07:34:18.5110343Z 2025-09-07T07:34:18.5110439Z Is the current job unstable? False 2025-09-07T07:34:18.5110622Z 2025-09-07T07:34:18.5110710Z Is keep-going label set? True 2025-09-07T07:34:18.5110873Z 2025-09-07T07:34:18.5110948Z Reenabled issues? 2025-09-07T07:34:18.5234929Z ##[group]Run echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-09-07T07:34:18.5235458Z echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-09-07T07:34:18.5241412Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:34:18.5241735Z env: 2025-09-07T07:34:18.5241914Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:18.5242135Z JOB_TIMEOUT: 720 2025-09-07T07:34:18.5242319Z ##[endgroup] 2025-09-07T07:34:18.5344391Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:34:18.5344867Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:34:18.5345274Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:34:18.5350931Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:34:18.5351254Z env: 2025-09-07T07:34:18.5351432Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:18.5351644Z ##[endgroup] 2025-09-07T07:34:18.5532325Z ##[group]Run set -x 2025-09-07T07:34:18.5532662Z set -x 2025-09-07T07:34:18.5532847Z  2025-09-07T07:34:18.5533055Z if [[ $TEST_CONFIG == 'multigpu' ]]; then 2025-09-07T07:34:18.5533391Z  TEST_COMMAND=.ci/pytorch/multigpu-test.sh 2025-09-07T07:34:18.5533731Z elif [[ $BUILD_ENVIRONMENT == *onnx* ]]; then 2025-09-07T07:34:18.5534036Z  TEST_COMMAND=.ci/onnx/test.sh 2025-09-07T07:34:18.5534282Z else 2025-09-07T07:34:18.5534489Z  TEST_COMMAND=.ci/pytorch/test.sh 2025-09-07T07:34:18.5534743Z fi 2025-09-07T07:34:18.5534906Z  2025-09-07T07:34:18.5535121Z # Leaving 1GB for the runner and other things 2025-09-07T07:34:18.5535813Z TOTAL_AVAILABLE_MEMORY_IN_GB=$(awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo) 2025-09-07T07:34:18.5536562Z # https://docs.docker.com/engine/containers/resource_constraints/#--memory-swap-details, the 3GB swap 2025-09-07T07:34:18.5537176Z # comes from https://github.com/pytorch/test-infra/pull/6058 2025-09-07T07:34:18.5537632Z TOTAL_MEMORY_WITH_SWAP=$(("${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}" + 3)) 2025-09-07T07:34:18.5537986Z  2025-09-07T07:34:18.5538206Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-09-07T07:34:18.5538499Z  SHM_OPTS= 2025-09-07T07:34:18.5538700Z  JENKINS_USER= 2025-09-07T07:34:18.5538997Z  # ensure that docker container cleanly exits in 12 hours 2025-09-07T07:34:18.5539410Z  # if for some reason cleanup action doesn't stop container 2025-09-07T07:34:18.5539750Z  # when job is cancelled 2025-09-07T07:34:18.5540019Z  DOCKER_SHELL_CMD="sleep 12h" 2025-09-07T07:34:18.5540265Z else 2025-09-07T07:34:18.5540471Z  SHM_OPTS="--shm-size=${SHM_SIZE}" 2025-09-07T07:34:18.5540759Z  JENKINS_USER="--user jenkins" 2025-09-07T07:34:18.5541027Z  DOCKER_SHELL_CMD= 2025-09-07T07:34:18.5541237Z fi 2025-09-07T07:34:18.5541401Z  2025-09-07T07:34:18.5541688Z # detached container should get cleaned up by teardown_ec2_linux 2025-09-07T07:34:18.5542147Z # TODO: Stop building test binaries as part of the build phase 2025-09-07T07:34:18.5542672Z # Used for GPU_FLAG, SHM_OPTS, JENKINS_USER and DOCKER_SHELL_CMD since that doesn't play nice 2025-09-07T07:34:18.5543135Z # shellcheck disable=SC2086,SC2090 2025-09-07T07:34:18.5543422Z container_name=$(docker run \ 2025-09-07T07:34:18.5543684Z  ${GPU_FLAG:-} \ 2025-09-07T07:34:18.5543937Z  ${SCCACHE_SERVER_PORT_DOCKER_FLAG:-} \ 2025-09-07T07:34:18.5544233Z  -e BUILD_ENVIRONMENT \ 2025-09-07T07:34:18.5544483Z  -e PR_NUMBER \ 2025-09-07T07:34:18.5544710Z  -e GITHUB_ACTIONS \ 2025-09-07T07:34:18.5544951Z  -e GITHUB_REPOSITORY \ 2025-09-07T07:34:18.5545203Z  -e GITHUB_WORKFLOW \ 2025-09-07T07:34:18.5545443Z  -e GITHUB_JOB \ 2025-09-07T07:34:18.5545669Z  -e GITHUB_RUN_ID \ 2025-09-07T07:34:18.5545906Z  -e GITHUB_RUN_NUMBER \ 2025-09-07T07:34:18.5546151Z  -e GITHUB_RUN_ATTEMPT \ 2025-09-07T07:34:18.5546395Z  -e JOB_ID \ 2025-09-07T07:34:18.5546602Z  -e JOB_NAME \ 2025-09-07T07:34:18.5546815Z  -e BASE_SHA \ 2025-09-07T07:34:18.5547021Z  -e BRANCH \ 2025-09-07T07:34:18.5547223Z  -e SHA1 \ 2025-09-07T07:34:18.5547429Z  -e AWS_DEFAULT_REGION \ 2025-09-07T07:34:18.5547676Z  -e IN_WHEEL_TEST \ 2025-09-07T07:34:18.5547902Z  -e SHARD_NUMBER \ 2025-09-07T07:34:18.5548135Z  -e TEST_CONFIG \ 2025-09-07T07:34:18.5548362Z  -e NUM_TEST_SHARDS \ 2025-09-07T07:34:18.5548604Z  -e REENABLED_ISSUES \ 2025-09-07T07:34:18.5548854Z  -e CONTINUE_THROUGH_ERROR \ 2025-09-07T07:34:18.5549294Z  -e VERBOSE_TEST_LOGS \ 2025-09-07T07:34:18.5549560Z  -e TEST_SHOWLOCALS \ 2025-09-07T07:34:18.5549803Z  -e NO_TEST_TIMEOUT \ 2025-09-07T07:34:18.5550027Z  -e NO_TD \ 2025-09-07T07:34:18.5550236Z  -e TD_DISTRIBUTED \ 2025-09-07T07:34:18.5550470Z  -e PR_LABELS \ 2025-09-07T07:34:18.5550716Z  -e MAX_JOBS="$(nproc --ignore=2)" \ 2025-09-07T07:34:18.5550992Z  -e SCCACHE_BUCKET \ 2025-09-07T07:34:18.5551228Z  -e SCCACHE_REGION \ 2025-09-07T07:34:18.5551458Z  -e XLA_CUDA \ 2025-09-07T07:34:18.5551704Z  -e XLA_CLANG_CACHE_S3_BUCKET_NAME \ 2025-09-07T07:34:18.5552013Z  -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK \ 2025-09-07T07:34:18.5552428Z  -e PYTORCH_TEST_RERUN_DISABLED_TESTS \ 2025-09-07T07:34:18.5552744Z  -e SKIP_SCCACHE_INITIALIZATION=1 \ 2025-09-07T07:34:18.5553034Z  -e HUGGING_FACE_HUB_TOKEN \ 2025-09-07T07:34:18.5553320Z  -e VLLM_TEST_HUGGING_FACE_TOKEN \ 2025-09-07T07:34:18.5553612Z  -e SCRIBE_GRAPHQL_ACCESS_TOKEN \ 2025-09-07T07:34:18.5553886Z  -e DASHBOARD_TAG \ 2025-09-07T07:34:18.5554126Z  -e ARTIFACTS_FILE_SUFFIX \ 2025-09-07T07:34:18.5554437Z  --memory="${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}g" \ 2025-09-07T07:34:18.5554789Z  --memory-swap="${TOTAL_MEMORY_WITH_SWAP}g" \ 2025-09-07T07:34:18.5555142Z  --env-file="/tmp/github_env_${GITHUB_RUN_ID}" \ 2025-09-07T07:34:18.5555475Z  --security-opt seccomp=unconfined \ 2025-09-07T07:34:18.5555765Z  --cap-add=SYS_PTRACE \ 2025-09-07T07:34:18.5556007Z  --ipc=host \ 2025-09-07T07:34:18.5556221Z  ${SHM_OPTS} \ 2025-09-07T07:34:18.5556425Z  --tty \ 2025-09-07T07:34:18.5556621Z  --detach \ 2025-09-07T07:34:18.5556842Z  --name="${container_name}" \ 2025-09-07T07:34:18.5557097Z  ${JENKINS_USER} \ 2025-09-07T07:34:18.5557391Z  -v "${GITHUB_WORKSPACE}:/var/lib/jenkins/workspace" \ 2025-09-07T07:34:18.5557730Z  -w /var/lib/jenkins/workspace \ 2025-09-07T07:34:18.5558002Z  "${DOCKER_IMAGE}" \ 2025-09-07T07:34:18.5558234Z  ${DOCKER_SHELL_CMD} 2025-09-07T07:34:18.5558452Z ) 2025-09-07T07:34:18.5558691Z # Propagate download.pytorch.org IP to container 2025-09-07T07:34:18.5559262Z grep download.pytorch.org /etc/hosts | docker exec -i "${container_name}" sudo bash -c "/bin/cat >> /etc/hosts" 2025-09-07T07:34:18.5559875Z echo "DOCKER_CONTAINER_ID=${container_name}" >> "${GITHUB_ENV}" 2025-09-07T07:34:18.5560217Z  2025-09-07T07:34:18.5560435Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-09-07T07:34:18.5560942Z  docker exec -t "${container_name}" sh -c "python3 -m pip install -r .ci/docker/requirements-ci.txt" 2025-09-07T07:34:18.5561383Z fi 2025-09-07T07:34:18.5561548Z  2025-09-07T07:34:18.5561968Z docker exec -t "${container_name}" sh -c "python3 -m pip install $(echo dist/*.whl)[opt-einsum] && ${TEST_COMMAND}" 2025-09-07T07:34:18.5568067Z shell: /usr/bin/bash -e {0} 2025-09-07T07:34:18.5568287Z env: 2025-09-07T07:34:18.5568466Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:34:18.5568734Z BUILD_ENVIRONMENT: linux-jammy-aarch64-py3.10 2025-09-07T07:34:18.5569015Z PR_NUMBER: 2025-09-07T07:34:18.5569211Z GITHUB_REPOSITORY: pytorch/pytorch 2025-09-07T07:34:18.5569509Z GITHUB_WORKFLOW: inductor-perf-nightly-aarch64 2025-09-07T07:34:18.5569791Z GITHUB_JOB: test 2025-09-07T07:34:18.5569986Z GITHUB_RUN_ID: 17525321311 2025-09-07T07:34:18.5570208Z GITHUB_RUN_NUMBER: 706 2025-09-07T07:34:18.5570426Z GITHUB_RUN_ATTEMPT: 1 2025-09-07T07:34:18.5570625Z JOB_ID: 49775354812 2025-09-07T07:34:18.5571119Z JOB_NAME: linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal) 2025-09-07T07:34:18.5571653Z BRANCH: main 2025-09-07T07:34:18.5572059Z SHA1: 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:34:18.5572394Z BASE_SHA: 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:34:18.5572726Z TEST_CONFIG: inductor_torchbench_perf_cpu_aarch64 2025-09-07T07:34:18.5573011Z SHARD_NUMBER: 8 2025-09-07T07:34:18.5573198Z NUM_TEST_SHARDS: 15 2025-09-07T07:34:18.5573392Z REENABLED_ISSUES: 2025-09-07T07:34:18.5573599Z CONTINUE_THROUGH_ERROR: True 2025-09-07T07:34:18.5573833Z VERBOSE_TEST_LOGS: False 2025-09-07T07:34:18.5574052Z TEST_SHOWLOCALS: False 2025-09-07T07:34:18.5574260Z NO_TEST_TIMEOUT: False 2025-09-07T07:34:18.5574458Z NO_TD: False 2025-09-07T07:34:18.5574640Z TD_DISTRIBUTED: False 2025-09-07T07:34:18.5574898Z SCCACHE_BUCKET: ossci-compiler-cache-circleci-v2 2025-09-07T07:34:18.5575349Z SCCACHE_REGION: us-east-1 2025-09-07T07:34:18.5575564Z SHM_SIZE: 1g 2025-09-07T07:34:18.5576326Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:34:18.5577137Z XLA_CUDA: 2025-09-07T07:34:18.5577438Z XLA_CLANG_CACHE_S3_BUCKET_NAME: ossci-compiler-clang-cache-circleci-xla 2025-09-07T07:34:18.5577823Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK: 0 2025-09-07T07:34:18.5578094Z PYTORCH_TEST_RERUN_DISABLED_TESTS: 0 2025-09-07T07:34:18.5578663Z DASHBOARD_TAG: training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true 2025-09-07T07:34:18.5579421Z VLLM_TEST_HUGGING_FACE_TOKEN: *** 2025-09-07T07:34:18.5579770Z HUGGING_FACE_HUB_TOKEN: *** 2025-09-07T07:34:18.5580119Z SCRIBE_GRAPHQL_ACCESS_TOKEN: *** 2025-09-07T07:34:18.5580591Z ARTIFACTS_FILE_SUFFIX: test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812 2025-09-07T07:34:18.5581074Z ##[endgroup] 2025-09-07T07:34:18.5602405Z + [[ inductor_torchbench_perf_cpu_aarch64 == \m\u\l\t\i\g\p\u ]] 2025-09-07T07:34:18.5602798Z + [[ linux-jammy-aarch64-py3.10 == *onnx* ]] 2025-09-07T07:34:18.5603090Z + TEST_COMMAND=.ci/pytorch/test.sh 2025-09-07T07:34:18.5605329Z ++ awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo 2025-09-07T07:34:18.5623776Z + TOTAL_AVAILABLE_MEMORY_IN_GB='250.189 ' 2025-09-07T07:34:18.5624058Z + TOTAL_MEMORY_WITH_SWAP=253 2025-09-07T07:34:18.5624336Z + [[ linux-jammy-aarch64-py3.10 == *\s\3\9\0\x* ]] 2025-09-07T07:34:18.5624625Z + SHM_OPTS=--shm-size=1g 2025-09-07T07:34:18.5624848Z + JENKINS_USER='--user jenkins' 2025-09-07T07:34:18.5625077Z + DOCKER_SHELL_CMD= 2025-09-07T07:34:18.5631440Z +++ nproc --ignore=2 2025-09-07T07:34:18.6334557Z ++ docker run -e BUILD_ENVIRONMENT -e PR_NUMBER -e GITHUB_ACTIONS -e GITHUB_REPOSITORY -e GITHUB_WORKFLOW -e GITHUB_JOB -e GITHUB_RUN_ID -e GITHUB_RUN_NUMBER -e GITHUB_RUN_ATTEMPT -e JOB_ID -e JOB_NAME -e BASE_SHA -e BRANCH -e SHA1 -e AWS_DEFAULT_REGION -e IN_WHEEL_TEST -e SHARD_NUMBER -e TEST_CONFIG -e NUM_TEST_SHARDS -e REENABLED_ISSUES -e CONTINUE_THROUGH_ERROR -e VERBOSE_TEST_LOGS -e TEST_SHOWLOCALS -e NO_TEST_TIMEOUT -e NO_TD -e TD_DISTRIBUTED -e PR_LABELS -e MAX_JOBS=62 -e SCCACHE_BUCKET -e SCCACHE_REGION -e XLA_CUDA -e XLA_CLANG_CACHE_S3_BUCKET_NAME -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK -e PYTORCH_TEST_RERUN_DISABLED_TESTS -e SKIP_SCCACHE_INITIALIZATION=1 -e HUGGING_FACE_HUB_TOKEN -e VLLM_TEST_HUGGING_FACE_TOKEN -e SCRIBE_GRAPHQL_ACCESS_TOKEN -e DASHBOARD_TAG -e ARTIFACTS_FILE_SUFFIX --memory=250g --memory-swap=253g --env-file=/tmp/github_env_17525321311 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --ipc=host --shm-size=1g --tty --detach --name= --user jenkins -v /home/ec2-user/actions-runner/_work/pytorch/pytorch:/var/lib/jenkins/workspace -w /var/lib/jenkins/workspace 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:35:20.5894602Z + container_name=4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T07:35:20.5896743Z + grep download.pytorch.org /etc/hosts 2025-09-07T07:35:20.5899038Z + docker exec -i 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d sudo bash -c '/bin/cat >> /etc/hosts' 2025-09-07T07:35:20.7399136Z + echo DOCKER_CONTAINER_ID=4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T07:35:20.7400413Z + [[ linux-jammy-aarch64-py3.10 == *\s\3\9\0\x* ]] 2025-09-07T07:35:20.7403979Z ++ echo dist/torch-2.9.0a0+git93fb23d-cp310-cp310-linux_aarch64.whl 2025-09-07T07:35:20.7405969Z + docker exec -t 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d sh -c 'python3 -m pip install dist/torch-2.9.0a0+git93fb23d-cp310-cp310-linux_aarch64.whl[opt-einsum] && .ci/pytorch/test.sh' 2025-09-07T07:35:21.1486246Z Processing ./dist/torch-2.9.0a0+git93fb23d-cp310-cp310-linux_aarch64.whl (from torch==2.9.0a0+git93fb23d) 2025-09-07T07:35:21.3426487Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (3.19.1) 2025-09-07T07:35:21.3431934Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (4.15.0) 2025-09-07T07:35:21.3437186Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (1.13.3) 2025-09-07T07:35:21.3442453Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (2.8.8) 2025-09-07T07:35:21.3446433Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (3.1.6) 2025-09-07T07:35:21.3451633Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (2025.3.0) 2025-09-07T07:35:21.3466705Z Requirement already satisfied: opt-einsum>=3.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (3.3.0) 2025-09-07T07:35:21.3924772Z Requirement already satisfied: numpy>=1.7 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from opt-einsum>=3.3->torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (1.22.4) 2025-09-07T07:35:21.3947832Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from sympy>=1.13.3->torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (1.3.0) 2025-09-07T07:35:21.3991157Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from jinja2->torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (3.0.2) 2025-09-07T07:35:22.4579750Z Installing collected packages: torch 2025-09-07T07:35:31.3423793Z ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts. 2025-09-07T07:35:31.3424634Z dall-e 0.1 requires torchvision, which is not installed. 2025-09-07T07:35:31.3425016Z effdet 0.4.1 requires torchvision, which is not installed. 2025-09-07T07:35:31.3425449Z python-doctr 1.0.0 requires torchvision>=0.15.0, which is not installed. 2025-09-07T07:35:31.3425984Z pytorch-labs-segment-anything-fast 0.2 requires torchao, which is not installed. 2025-09-07T07:35:31.3426648Z pytorch-labs-segment-anything-fast 0.2 requires torchvision>=0.17.0.dev20231026, which is not installed. 2025-09-07T07:35:31.3427344Z timm 1.0.14 requires torchvision, which is not installed. 2025-09-07T07:35:31.3427796Z Successfully installed torch-2.9.0a0+git93fb23d 2025-09-07T07:35:31.4152302Z + export TERM=vt100 2025-09-07T07:35:31.4152509Z + TERM=vt100 2025-09-07T07:35:31.4155019Z ++ dirname .ci/pytorch/test.sh 2025-09-07T07:35:31.4161798Z + source .ci/pytorch/common.sh 2025-09-07T07:35:31.4164633Z +++ dirname .ci/pytorch/common.sh 2025-09-07T07:35:31.4171074Z ++ source .ci/pytorch/common_utils.sh 2025-09-07T07:35:31.4171631Z +++ declare -f -t trap_add 2025-09-07T07:35:31.4175348Z ++ set -ex -o pipefail 2025-09-07T07:35:31.4175599Z ++ [[ linux-jammy-aarch64-py3.10 == *rocm* ]] 2025-09-07T07:35:31.4175879Z ++ BUILD_TEST_LIBTORCH=0 2025-09-07T07:35:31.4178464Z ++ dirname .ci/pytorch/test.sh 2025-09-07T07:35:31.4184527Z + source .ci/pytorch/common-build.sh 2025-09-07T07:35:31.4185801Z ++ [[ linux-jammy-aarch64-py3.10 != *win-* ]] 2025-09-07T07:35:31.4190944Z ++++ dirname .ci/pytorch/common-build.sh 2025-09-07T07:35:31.4197190Z +++ cd .ci/pytorch 2025-09-07T07:35:31.4197664Z +++ pwd -P 2025-09-07T07:35:31.4199342Z ++ script_dir=/var/lib/jenkins/workspace/.ci/pytorch 2025-09-07T07:35:31.4199920Z ++ [[ linux-jammy-aarch64-py3.10 == *-pch* ]] 2025-09-07T07:35:31.4200190Z ++ which sccache 2025-09-07T07:35:31.4216004Z ++ [[ -z ossci-compiler-cache-circleci-v2 ]] 2025-09-07T07:35:31.4216312Z ++ sccache --stop-server 2025-09-07T07:35:31.4238655Z ++ true 2025-09-07T07:35:31.4238852Z ++ rm -f /var/lib/jenkins/sccache_error.log 2025-09-07T07:35:31.4246544Z ++ trap_add sccache_epilogue EXIT 2025-09-07T07:35:31.4246808Z ++ trap_add_cmd=sccache_epilogue 2025-09-07T07:35:31.4247036Z ++ shift 2025-09-07T07:35:31.4247224Z ++ for trap_add_name in "$@" 2025-09-07T07:35:31.4252666Z ++++ trap -p EXIT 2025-09-07T07:35:31.4254238Z +++ eval 'extract_trap_cmd ' 2025-09-07T07:35:31.4254487Z ++++ extract_trap_cmd 2025-09-07T07:35:31.4254693Z ++++ printf '%s\n' '' 2025-09-07T07:35:31.4255235Z +++ printf '%s\n' sccache_epilogue 2025-09-07T07:35:31.4256473Z ++ trap -- ' 2025-09-07T07:35:31.4256669Z sccache_epilogue' EXIT 2025-09-07T07:35:31.4256882Z ++ [[ -n 1 ]] 2025-09-07T07:35:31.4257234Z ++ echo 'Skipping sccache server initialization, setting environment variables' 2025-09-07T07:35:31.4257753Z Skipping sccache server initialization, setting environment variables 2025-09-07T07:35:31.4258137Z ++ export SCCACHE_IDLE_TIMEOUT=0 2025-09-07T07:35:31.4258390Z ++ SCCACHE_IDLE_TIMEOUT=0 2025-09-07T07:35:31.4258692Z ++ export SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-09-07T07:35:31.4259072Z ++ SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-09-07T07:35:31.4259413Z ++ export RUST_LOG=sccache::server=error 2025-09-07T07:35:31.4259686Z ++ RUST_LOG=sccache::server=error 2025-09-07T07:35:31.4259929Z ++ sccache --zero-stats 2025-09-07T07:35:31.5766021Z Statistics zeroed. 2025-09-07T07:35:31.5773060Z ++ which ccache 2025-09-07T07:35:31.6354522Z + [[ linux-jammy-aarch64-py3.10 != *rocm* ]] 2025-09-07T07:35:31.6354849Z + [[ linux-jammy-aarch64-py3.10 != *s390x* ]] 2025-09-07T07:35:31.6355143Z + [[ -d /var/lib/jenkins/workspace ]] 2025-09-07T07:35:31.6356635Z ++ stat -c %u /var/lib/jenkins/workspace 2025-09-07T07:35:31.6366404Z + WORKSPACE_ORIGINAL_OWNER_ID=1000 2025-09-07T07:35:31.6366674Z + trap_add cleanup_workspace EXIT 2025-09-07T07:35:31.6366933Z + trap_add_cmd=cleanup_workspace 2025-09-07T07:35:31.6367159Z + shift 2025-09-07T07:35:31.6367352Z + for trap_add_name in "$@" 2025-09-07T07:35:31.6372549Z +++ trap -p EXIT 2025-09-07T07:35:31.6374451Z ++ eval 'extract_trap_cmd trap -- '\'' 2025-09-07T07:35:31.6374717Z sccache_epilogue'\'' EXIT' 2025-09-07T07:35:31.6374963Z +++ extract_trap_cmd trap -- ' 2025-09-07T07:35:31.6375197Z sccache_epilogue' EXIT 2025-09-07T07:35:31.6375402Z +++ printf '%s\n' ' 2025-09-07T07:35:31.6375588Z sccache_epilogue' 2025-09-07T07:35:31.6375814Z ++ printf '%s\n' cleanup_workspace 2025-09-07T07:35:31.6376681Z + trap -- ' 2025-09-07T07:35:31.6376854Z sccache_epilogue 2025-09-07T07:35:31.6377048Z cleanup_workspace' EXIT 2025-09-07T07:35:31.6377309Z + sudo chown -R jenkins /var/lib/jenkins/workspace 2025-09-07T07:35:32.1854874Z + git config --global --add safe.directory /var/lib/jenkins/workspace 2025-09-07T07:35:32.1870548Z + echo 'Environment variables:' 2025-09-07T07:35:32.1870799Z Environment variables: 2025-09-07T07:35:32.1870996Z + env 2025-09-07T07:35:32.1877600Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-09-07T07:35:32.1878211Z CONTINUE_THROUGH_ERROR=True 2025-09-07T07:35:32.1878519Z BUILD_ENVIRONMENT=linux-jammy-aarch64-py3.10 2025-09-07T07:35:32.1879042Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-09-07T07:35:32.1879327Z HOSTNAME=4646a3beb296 2025-09-07T07:35:32.1879816Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_710a8386-9d54-4858-99a3-5883d1c550af 2025-09-07T07:35:32.1880355Z GITHUB_ACTION=__run_2 2025-09-07T07:35:32.1880577Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-09-07T07:35:32.1880823Z GITHUB_RUN_NUMBER=706 2025-09-07T07:35:32.1881070Z TEST_CONFIG=inductor_torchbench_perf_cpu_aarch64 2025-09-07T07:35:32.1881374Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-09-07T07:35:32.1881851Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-09-07T07:35:32.1882115Z SCCACHE_IDLE_TIMEOUT=0 2025-09-07T07:35:32.1882456Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-09-07T07:35:32.1882715Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-09-07T07:35:32.1882980Z GITHUB_REF_TYPE=branch 2025-09-07T07:35:32.1883228Z BASE_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:35:32.1883501Z XLA_CUDA= 2025-09-07T07:35:32.1883682Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-09-07T07:35:32.1884034Z HUGGING_FACE_HUB_TOKEN=*** 2025-09-07T07:35:32.1884442Z *** 2025-09-07T07:35:32.1884617Z GITHUB_REPOSITORY_ID=65600975 2025-09-07T07:35:32.1884845Z GITHUB_ACTIONS=true 2025-09-07T07:35:32.1885090Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-09-07T07:35:32.1885421Z SHA1=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:35:32.1885729Z GITHUB_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:35:32.1886299Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/inductor-perf-test-nightly-aarch64.yml@refs/heads/main 2025-09-07T07:35:32.1886824Z UCC_HOME=/usr 2025-09-07T07:35:32.1887010Z VERBOSE_TEST_LOGS=False 2025-09-07T07:35:32.1887228Z GITHUB_REF=refs/heads/main 2025-09-07T07:35:32.1887441Z SHARD_NUMBER=8 2025-09-07T07:35:32.1887629Z GITHUB_REF_PROTECTED=true 2025-09-07T07:35:32.1887851Z HOME=/var/lib/jenkins 2025-09-07T07:35:32.1888084Z GITHUB_API_URL=https://api.github.com 2025-09-07T07:35:32.1888362Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-09-07T07:35:32.1888599Z UCX_COMMIT= 2025-09-07T07:35:32.1888769Z USE_SYSTEM_NCCL=1 2025-09-07T07:35:32.1888961Z NUM_TEST_SHARDS=15 2025-09-07T07:35:32.1889150Z UCX_HOME=/usr 2025-09-07T07:35:32.1889628Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_710a8386-9d54-4858-99a3-5883d1c550af 2025-09-07T07:35:32.1890462Z JOB_NAME=linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal) 2025-09-07T07:35:32.1891280Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_710a8386-9d54-4858-99a3-5883d1c550af 2025-09-07T07:35:32.1891975Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-09-07T07:35:32.1892401Z GITHUB_EVENT_NAME=schedule 2025-09-07T07:35:32.1892951Z DASHBOARD_TAG=training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true 2025-09-07T07:35:32.1893524Z GITHUB_RUN_ID=17525321311 2025-09-07T07:35:32.1893741Z INSTALLED_OPENBLAS=yes 2025-09-07T07:35:32.1894276Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_710a8386-9d54-4858-99a3-5883d1c550af 2025-09-07T07:35:32.1894858Z GITHUB_ACTOR=pytorchmergebot 2025-09-07T07:35:32.1895083Z PR_NUMBER= 2025-09-07T07:35:32.1895249Z DESIRED_CUDA= 2025-09-07T07:35:32.1895432Z GITHUB_RUN_ATTEMPT=1 2025-09-07T07:35:32.1895642Z ANACONDA_PYTHON_VERSION=3.10 2025-09-07T07:35:32.1895926Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-09-07T07:35:32.1896218Z TERM=vt100 2025-09-07T07:35:32.1896387Z INSTALLED_VISION=yes 2025-09-07T07:35:32.1896587Z BRANCH=main 2025-09-07T07:35:32.1896769Z SCCACHE_REGION=us-east-1 2025-09-07T07:35:32.1896994Z OPENSSL_ROOT_DIR=/opt/openssl 2025-09-07T07:35:32.1897229Z CUDA_PATH=/usr/local/cuda 2025-09-07T07:35:32.1897835Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-09-07T07:35:32.1898375Z GITHUB_SERVER_URL=https://github.com 2025-09-07T07:35:32.1898624Z UCC_COMMIT= 2025-09-07T07:35:32.1898794Z REENABLED_ISSUES= 2025-09-07T07:35:32.1898970Z DOCS= 2025-09-07T07:35:32.1899123Z SHLVL=1 2025-09-07T07:35:32.1899280Z MAX_JOBS=62 2025-09-07T07:35:32.1899455Z GITHUB_ACTOR_ID=97764156 2025-09-07T07:35:32.1899738Z GITHUB_WORKFLOW_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:35:32.1900062Z GITHUB_REF_NAME=main 2025-09-07T07:35:32.1900384Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-09-07T07:35:32.1900745Z GITHUB_JOB=test 2025-09-07T07:35:32.1900927Z NO_TEST_TIMEOUT=False 2025-09-07T07:35:32.1901240Z TD_DISTRIBUTED=False 2025-09-07T07:35:32.1901462Z GITHUB_REPOSITORY=pytorch/pytorch 2025-09-07T07:35:32.1901717Z GITHUB_RETENTION_DAYS=90 2025-09-07T07:35:32.1901932Z OPENSSL_DIR=/opt/openssl 2025-09-07T07:35:32.1902152Z GITHUB_ACTION_REPOSITORY= 2025-09-07T07:35:32.1902828Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-09-07T07:35:32.1903521Z GITHUB_BASE_REF= 2025-09-07T07:35:32.1903705Z INSTALLED_ACL=yes 2025-09-07T07:35:32.1904129Z ARTIFACTS_FILE_SUFFIX=test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812 2025-09-07T07:35:32.1904600Z CI=true 2025-09-07T07:35:32.1904780Z GITHUB_REPOSITORY_OWNER=pytorch 2025-09-07T07:35:32.1905051Z RUST_LOG=sccache::server=error 2025-09-07T07:35:32.1905284Z JOB_ID=49775354812 2025-09-07T07:35:32.1905470Z GITHUB_HEAD_REF= 2025-09-07T07:35:32.1905656Z GITHUB_ACTION_REF= 2025-09-07T07:35:32.1905894Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-09-07T07:35:32.1906200Z TEST_SHOWLOCALS=False 2025-09-07T07:35:32.1906452Z GITHUB_WORKFLOW=inductor-perf-nightly-aarch64 2025-09-07T07:35:32.1906752Z DEBIAN_FRONTEND=noninteractive 2025-09-07T07:35:32.1907280Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_710a8386-9d54-4858-99a3-5883d1c550af 2025-09-07T07:35:32.1907815Z NO_TD=False 2025-09-07T07:35:32.1908003Z SKIP_SCCACHE_INITIALIZATION=1 2025-09-07T07:35:32.1908257Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-09-07T07:35:32.1908506Z _=/usr/bin/env 2025-09-07T07:35:32.1908774Z ++ python -c 'import site; print(site.getsitepackages()[0])' 2025-09-07T07:35:32.2751604Z + TORCH_INSTALL_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch 2025-09-07T07:35:32.2752141Z + TORCH_BIN_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-09-07T07:35:32.2752655Z + TORCH_LIB_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib 2025-09-07T07:35:32.2753185Z + TORCH_TEST_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/test 2025-09-07T07:35:32.2753565Z + BUILD_DIR=build 2025-09-07T07:35:32.2753766Z + BUILD_RENAMED_DIR=build_renamed 2025-09-07T07:35:32.2754014Z + BUILD_BIN_DIR=build/bin 2025-09-07T07:35:32.2754232Z + SHARD_NUMBER=8 2025-09-07T07:35:32.2754422Z + NUM_TEST_SHARDS=15 2025-09-07T07:35:32.2754638Z + export TORCH_SERIALIZATION_DEBUG=1 2025-09-07T07:35:32.2754907Z + TORCH_SERIALIZATION_DEBUG=1 2025-09-07T07:35:32.2755147Z + export VALGRIND=ON 2025-09-07T07:35:32.2755346Z + VALGRIND=ON 2025-09-07T07:35:32.2755560Z + [[ linux-jammy-aarch64-py3.10 == *clang9* ]] 2025-09-07T07:35:32.2755861Z + [[ linux-jammy-aarch64-py3.10 == *xpu* ]] 2025-09-07T07:35:32.2756121Z + detect_cuda_arch 2025-09-07T07:35:32.2756340Z + [[ linux-jammy-aarch64-py3.10 == *cuda* ]] 2025-09-07T07:35:32.2756636Z + [[ linux-jammy-aarch64-py3.10 == *s390x* ]] 2025-09-07T07:35:32.2756893Z + [[ 0 == \1 ]] 2025-09-07T07:35:32.2757066Z + [[ True == \1 ]] 2025-09-07T07:35:32.2757291Z + [[ linux-jammy-aarch64-py3.10 != *bazel* ]] 2025-09-07T07:35:32.2757586Z ++ realpath build/custom_test_artifacts 2025-09-07T07:35:32.2763162Z + CUSTOM_TEST_ARTIFACT_BUILD_DIR=/var/lib/jenkins/workspace/build/custom_test_artifacts 2025-09-07T07:35:32.2763593Z + [[ -n '' ]] 2025-09-07T07:35:32.2764039Z + echo 'Environment variables' 2025-09-07T07:35:32.2764291Z Environment variables 2025-09-07T07:35:32.2764481Z + env 2025-09-07T07:35:32.3107107Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-09-07T07:35:32.3107486Z CONTINUE_THROUGH_ERROR=True 2025-09-07T07:35:32.3107756Z BUILD_ENVIRONMENT=linux-jammy-aarch64-py3.10 2025-09-07T07:35:32.3108216Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-09-07T07:35:32.3108458Z HOSTNAME=4646a3beb296 2025-09-07T07:35:32.3108949Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_710a8386-9d54-4858-99a3-5883d1c550af 2025-09-07T07:35:32.3109486Z GITHUB_ACTION=__run_2 2025-09-07T07:35:32.3109705Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-09-07T07:35:32.3110204Z GITHUB_RUN_NUMBER=706 2025-09-07T07:35:32.3110454Z TEST_CONFIG=inductor_torchbench_perf_cpu_aarch64 2025-09-07T07:35:32.3110758Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-09-07T07:35:32.3111030Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-09-07T07:35:32.3111296Z SCCACHE_IDLE_TIMEOUT=0 2025-09-07T07:35:32.3111638Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-09-07T07:35:32.3111900Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-09-07T07:35:32.3112165Z GITHUB_REF_TYPE=branch 2025-09-07T07:35:32.3112411Z BASE_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:35:32.3112685Z XLA_CUDA= 2025-09-07T07:35:32.3112866Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-09-07T07:35:32.3113286Z HUGGING_FACE_HUB_TOKEN=*** 2025-09-07T07:35:32.3113565Z *** 2025-09-07T07:35:32.3113742Z GITHUB_REPOSITORY_ID=65600975 2025-09-07T07:35:32.3113966Z GITHUB_ACTIONS=true 2025-09-07T07:35:32.3114214Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-09-07T07:35:32.3114545Z SHA1=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:35:32.3114874Z GITHUB_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:35:32.3115443Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/inductor-perf-test-nightly-aarch64.yml@refs/heads/main 2025-09-07T07:35:32.3115964Z UCC_HOME=/usr 2025-09-07T07:35:32.3116166Z TORCH_SERIALIZATION_DEBUG=1 2025-09-07T07:35:32.3116396Z VERBOSE_TEST_LOGS=False 2025-09-07T07:35:32.3116606Z GITHUB_REF=refs/heads/main 2025-09-07T07:35:32.3116823Z SHARD_NUMBER=8 2025-09-07T07:35:32.3117011Z GITHUB_REF_PROTECTED=true 2025-09-07T07:35:32.3117227Z HOME=/var/lib/jenkins 2025-09-07T07:35:32.3117451Z GITHUB_API_URL=https://api.github.com 2025-09-07T07:35:32.3117728Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-09-07T07:35:32.3117967Z UCX_COMMIT= 2025-09-07T07:35:32.3118138Z USE_SYSTEM_NCCL=1 2025-09-07T07:35:32.3118323Z NUM_TEST_SHARDS=15 2025-09-07T07:35:32.3118508Z UCX_HOME=/usr 2025-09-07T07:35:32.3118986Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_710a8386-9d54-4858-99a3-5883d1c550af 2025-09-07T07:35:32.3119829Z JOB_NAME=linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal) 2025-09-07T07:35:32.3120643Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_710a8386-9d54-4858-99a3-5883d1c550af 2025-09-07T07:35:32.3121332Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-09-07T07:35:32.3121765Z GITHUB_EVENT_NAME=schedule 2025-09-07T07:35:32.3122323Z DASHBOARD_TAG=training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true 2025-09-07T07:35:32.3122892Z GITHUB_RUN_ID=17525321311 2025-09-07T07:35:32.3123104Z INSTALLED_OPENBLAS=yes 2025-09-07T07:35:32.3123641Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_710a8386-9d54-4858-99a3-5883d1c550af 2025-09-07T07:35:32.3124227Z GITHUB_ACTOR=pytorchmergebot 2025-09-07T07:35:32.3124452Z PR_NUMBER= 2025-09-07T07:35:32.3124623Z DESIRED_CUDA= 2025-09-07T07:35:32.3124806Z GITHUB_RUN_ATTEMPT=1 2025-09-07T07:35:32.3125002Z VALGRIND=ON 2025-09-07T07:35:32.3125191Z ANACONDA_PYTHON_VERSION=3.10 2025-09-07T07:35:32.3125471Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-09-07T07:35:32.3125978Z TERM=vt100 2025-09-07T07:35:32.3126177Z INSTALLED_VISION=yes 2025-09-07T07:35:32.3126374Z BRANCH=main 2025-09-07T07:35:32.3126548Z SCCACHE_REGION=us-east-1 2025-09-07T07:35:32.3126772Z OPENSSL_ROOT_DIR=/opt/openssl 2025-09-07T07:35:32.3127007Z CUDA_PATH=/usr/local/cuda 2025-09-07T07:35:32.3127452Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-09-07T07:35:32.3127946Z GITHUB_SERVER_URL=https://github.com 2025-09-07T07:35:32.3128192Z UCC_COMMIT= 2025-09-07T07:35:32.3128362Z REENABLED_ISSUES= 2025-09-07T07:35:32.3128538Z DOCS= 2025-09-07T07:35:32.3128687Z SHLVL=1 2025-09-07T07:35:32.3128844Z MAX_JOBS=62 2025-09-07T07:35:32.3129020Z GITHUB_ACTOR_ID=97764156 2025-09-07T07:35:32.3129414Z GITHUB_WORKFLOW_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:35:32.3129733Z GITHUB_REF_NAME=main 2025-09-07T07:35:32.3130056Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-09-07T07:35:32.3130414Z GITHUB_JOB=test 2025-09-07T07:35:32.3130602Z NO_TEST_TIMEOUT=False 2025-09-07T07:35:32.3130802Z TD_DISTRIBUTED=False 2025-09-07T07:35:32.3131022Z GITHUB_REPOSITORY=pytorch/pytorch 2025-09-07T07:35:32.3131275Z GITHUB_RETENTION_DAYS=90 2025-09-07T07:35:32.3131493Z OPENSSL_DIR=/opt/openssl 2025-09-07T07:35:32.3131707Z GITHUB_ACTION_REPOSITORY= 2025-09-07T07:35:32.3132376Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-09-07T07:35:32.3133062Z GITHUB_BASE_REF= 2025-09-07T07:35:32.3133249Z INSTALLED_ACL=yes 2025-09-07T07:35:32.3133667Z ARTIFACTS_FILE_SUFFIX=test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812 2025-09-07T07:35:32.3134141Z CI=true 2025-09-07T07:35:32.3134321Z GITHUB_REPOSITORY_OWNER=pytorch 2025-09-07T07:35:32.3134585Z RUST_LOG=sccache::server=error 2025-09-07T07:35:32.3134804Z JOB_ID=49775354812 2025-09-07T07:35:32.3134989Z GITHUB_HEAD_REF= 2025-09-07T07:35:32.3135180Z GITHUB_ACTION_REF= 2025-09-07T07:35:32.3135425Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-09-07T07:35:32.3135718Z TEST_SHOWLOCALS=False 2025-09-07T07:35:32.3135967Z GITHUB_WORKFLOW=inductor-perf-nightly-aarch64 2025-09-07T07:35:32.3136266Z DEBIAN_FRONTEND=noninteractive 2025-09-07T07:35:32.3136797Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_710a8386-9d54-4858-99a3-5883d1c550af 2025-09-07T07:35:32.3137329Z NO_TD=False 2025-09-07T07:35:32.3137520Z SKIP_SCCACHE_INITIALIZATION=1 2025-09-07T07:35:32.3137777Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-09-07T07:35:32.3138030Z _=/usr/bin/env 2025-09-07T07:35:32.3138209Z + echo 'Testing pytorch' 2025-09-07T07:35:32.3138418Z Testing pytorch 2025-09-07T07:35:32.3138615Z + export LANG=C.UTF-8 2025-09-07T07:35:32.3138808Z + LANG=C.UTF-8 2025-09-07T07:35:32.3138996Z + PR_NUMBER= 2025-09-07T07:35:32.3139253Z + [[ inductor_torchbench_perf_cpu_aarch64 == \d\e\f\a\u\l\t ]] 2025-09-07T07:35:32.3139679Z + [[ inductor_torchbench_perf_cpu_aarch64 == \d\i\s\t\r\i\b\u\t\e\d ]] 2025-09-07T07:35:32.3140087Z + [[ inductor_torchbench_perf_cpu_aarch64 == \s\l\o\w ]] 2025-09-07T07:35:32.3140443Z + [[ linux-jammy-aarch64-py3.10 == *slow-gradcheck* ]] 2025-09-07T07:35:32.3140765Z + [[ linux-jammy-aarch64-py3.10 == *cuda* ]] 2025-09-07T07:35:32.3141063Z + [[ linux-jammy-aarch64-py3.10 == *rocm* ]] 2025-09-07T07:35:32.3141354Z + [[ linux-jammy-aarch64-py3.10 == *xpu* ]] 2025-09-07T07:35:32.3141683Z + [[ inductor_torchbench_perf_cpu_aarch64 == *crossref* ]] 2025-09-07T07:35:32.3142012Z + [[ linux-jammy-aarch64-py3.10 == *rocm* ]] 2025-09-07T07:35:32.3142303Z + [[ linux-jammy-aarch64-py3.10 == *xpu* ]] 2025-09-07T07:35:32.3142598Z + [[ linux-jammy-aarch64-py3.10 != *-bazel-* ]] 2025-09-07T07:35:32.3142887Z + pip_install ninja==1.10.2 2025-09-07T07:35:32.3143193Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-09-07T07:35:32.3143598Z + python3 -m pip install --progress-bar off ninja==1.10.2 2025-09-07T07:35:33.8955558Z Collecting ninja==1.10.2 2025-09-07T07:35:33.9138868Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_17_aarch64.manylinux2014_aarch64.whl.metadata (5.0 kB) 2025-09-07T07:35:34.0071098Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (119 kB) 2025-09-07T07:35:35.1189757Z Installing collected packages: ninja 2025-09-07T07:35:35.1191009Z Attempting uninstall: ninja 2025-09-07T07:35:35.1203154Z Found existing installation: ninja 1.11.1.3 2025-09-07T07:35:35.1225640Z Uninstalling ninja-1.11.1.3: 2025-09-07T07:35:35.1286384Z Successfully uninstalled ninja-1.11.1.3 2025-09-07T07:35:35.1814496Z Successfully installed ninja-1.10.2 2025-09-07T07:35:35.2536378Z + export PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-09-07T07:35:35.2538165Z + PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-09-07T07:35:35.2539021Z + [[ linux-jammy-aarch64-py3.10 == *aarch64* ]] 2025-09-07T07:35:35.2539309Z + export VALGRIND=OFF 2025-09-07T07:35:35.2539508Z + VALGRIND=OFF 2025-09-07T07:35:35.2539722Z + [[ linux-jammy-aarch64-py3.10 == *asan* ]] 2025-09-07T07:35:35.2540026Z + [[ linux-jammy-aarch64-py3.10 == *-debug* ]] 2025-09-07T07:35:35.2540332Z + [[ linux-jammy-aarch64-py3.10 != *-bazel-* ]] 2025-09-07T07:35:35.2540775Z + echo 'We are not in debug mode: linux-jammy-aarch64-py3.10. Expect the assertion to pass' 2025-09-07T07:35:35.2541331Z We are not in debug mode: linux-jammy-aarch64-py3.10. Expect the assertion to pass 2025-09-07T07:35:35.2541723Z + cd test 2025-09-07T07:35:35.2542012Z + python -c 'import torch; torch._C._crash_if_debug_asserts_fail(424242)' 2025-09-07T07:35:35.5799047Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:35:35.5800295Z import pynvml # type: ignore[import] 2025-09-07T07:35:36.6622295Z + [[ inductor_torchbench_perf_cpu_aarch64 == \n\o\g\p\u\_\N\O\_\A\V\X\2 ]] 2025-09-07T07:35:36.6622791Z + [[ inductor_torchbench_perf_cpu_aarch64 == \n\o\g\p\u\_\A\V\X\5\1\2 ]] 2025-09-07T07:35:36.6623281Z + [[ inductor_torchbench_perf_cpu_aarch64 == \l\e\g\a\c\y\_\n\v\i\d\i\a\_\d\r\i\v\e\r ]] 2025-09-07T07:35:36.6625990Z + DYNAMO_BENCHMARK_FLAGS=() 2025-09-07T07:35:36.6626525Z + [[ inductor_torchbench_perf_cpu_aarch64 == *pr_time_benchmarks* ]] 2025-09-07T07:35:36.6626988Z + [[ inductor_torchbench_perf_cpu_aarch64 == *dynamo_eager* ]] 2025-09-07T07:35:36.6627390Z + [[ inductor_torchbench_perf_cpu_aarch64 == *aot_eager* ]] 2025-09-07T07:35:36.6627772Z + [[ inductor_torchbench_perf_cpu_aarch64 == *aot_inductor* ]] 2025-09-07T07:35:36.6628209Z + [[ inductor_torchbench_perf_cpu_aarch64 == *max_autotune_inductor* ]] 2025-09-07T07:35:36.6628618Z + [[ inductor_torchbench_perf_cpu_aarch64 == *inductor* ]] 2025-09-07T07:35:36.6628977Z + [[ inductor_torchbench_perf_cpu_aarch64 != *perf* ]] 2025-09-07T07:35:36.6629328Z + [[ inductor_torchbench_perf_cpu_aarch64 == *dynamic* ]] 2025-09-07T07:35:36.6629674Z + [[ inductor_torchbench_perf_cpu_aarch64 == *cpu* ]] 2025-09-07T07:35:36.6629994Z + DYNAMO_BENCHMARK_FLAGS+=(--device cpu) 2025-09-07T07:35:36.6652949Z + [[ linux-jammy-aarch64-py3.10 == *libtorch* ]] 2025-09-07T07:35:36.6653279Z + [[ linux-jammy-aarch64-py3.10 == *-bazel-* ]] 2025-09-07T07:35:36.6655741Z + cd test 2025-09-07T07:35:36.6656285Z + python -c 'import torch; print(torch.__config__.show())' 2025-09-07T07:35:36.9900895Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:35:36.9902136Z import pynvml # type: ignore[import] 2025-09-07T07:35:37.8609451Z PyTorch built with: 2025-09-07T07:35:37.8609668Z - GCC 11.4 2025-09-07T07:35:37.8609849Z - C++ Version: 201703 2025-09-07T07:35:37.8610177Z - Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-09-07T07:35:37.8610557Z - OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-09-07T07:35:37.8610846Z - LAPACK is enabled (usually provided by MKL) 2025-09-07T07:35:37.8611121Z - NNPACK is enabled 2025-09-07T07:35:37.8611348Z - CPU capability usage: SVE256 2025-09-07T07:35:37.8615511Z - Build settings: BLAS_INFO=open, BUILD_TYPE=Release, COMMIT_SHA=93fb23d6fae7c4e82c4239a1033e522088742634, CXX_COMPILER=/opt/cache/bin/c++, CXX_FLAGS= -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOCUPTI -DLIBKINETO_NOROCTRACER -DLIBKINETO_NOXPUPTI=ON -DUSE_PYTORCH_QNNPACK -DAT_BUILD_ARM_VEC256_WITH_SLEEF -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -DC10_NODEPRECATED -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=range-loop-construct -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-unknown-pragmas -Wno-unused-parameter -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=old-style-cast -faligned-new -Werror -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, LAPACK_INFO=open, TORCH_VERSION=2.9.0, USE_CUDA=OFF, USE_CUDNN=OFF, USE_CUSPARSELT=OFF, USE_EIGEN_FOR_BLAS=ON, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=OFF, USE_MKLDNN=1, USE_MPI=OFF, USE_NCCL=OFF, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, USE_ROCM_KERNEL_ASSERT=OFF, USE_XCCL=OFF, USE_XPU=OFF, 2025-09-07T07:35:37.8620014Z 2025-09-07T07:35:38.0764684Z + cd test 2025-09-07T07:35:38.0764994Z + python -c 'import torch; print(torch.__config__.parallel_info())' 2025-09-07T07:35:38.4025727Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:35:38.4026885Z import pynvml # type: ignore[import] 2025-09-07T07:35:39.2784759Z ATen/Parallel: 2025-09-07T07:35:39.2785076Z at::get_num_threads() : 64 2025-09-07T07:35:39.2785337Z at::get_num_interop_threads() : 64 2025-09-07T07:35:39.2785606Z OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-09-07T07:35:39.2785861Z omp_get_max_threads() : 64 2025-09-07T07:35:39.2786193Z Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-09-07T07:35:39.2786614Z std::thread::hardware_concurrency() : 64 2025-09-07T07:35:39.2786888Z Environment variables: 2025-09-07T07:35:39.2787108Z OMP_NUM_THREADS : [not set] 2025-09-07T07:35:39.2787346Z ATen parallel backend: OpenMP 2025-09-07T07:35:39.2787510Z 2025-09-07T07:35:39.4982021Z + [[ inductor_torchbench_perf_cpu_aarch64 == *numpy_2* ]] 2025-09-07T07:35:39.4982395Z + [[ linux-jammy-aarch64-py3.10 == *aarch64* ]] 2025-09-07T07:35:39.4982773Z + [[ inductor_torchbench_perf_cpu_aarch64 != *perf_cpu_aarch64* ]] 2025-09-07T07:35:39.4983180Z + [[ inductor_torchbench_perf_cpu_aarch64 == *backward* ]] 2025-09-07T07:35:39.4983553Z + [[ inductor_torchbench_perf_cpu_aarch64 == *xla* ]] 2025-09-07T07:35:39.4983890Z + [[ inductor_torchbench_perf_cpu_aarch64 == *vllm* ]] 2025-09-07T07:35:39.4984251Z + [[ inductor_torchbench_perf_cpu_aarch64 == *executorch* ]] 2025-09-07T07:35:39.4984656Z + [[ inductor_torchbench_perf_cpu_aarch64 == \j\i\t\_\l\e\g\a\c\y ]] 2025-09-07T07:35:39.4985034Z + [[ linux-jammy-aarch64-py3.10 == *libtorch* ]] 2025-09-07T07:35:39.4985396Z + [[ inductor_torchbench_perf_cpu_aarch64 == distributed ]] 2025-09-07T07:35:39.4985809Z + [[ inductor_torchbench_perf_cpu_aarch64 == *operator_benchmark* ]] 2025-09-07T07:35:39.4986548Z + [[ inductor_torchbench_perf_cpu_aarch64 == *inductor_distributed* ]] 2025-09-07T07:35:39.4987009Z + [[ inductor_torchbench_perf_cpu_aarch64 == *inductor-halide* ]] 2025-09-07T07:35:39.4987444Z + [[ inductor_torchbench_perf_cpu_aarch64 == *inductor-triton-cpu* ]] 2025-09-07T07:35:39.4987915Z + [[ inductor_torchbench_perf_cpu_aarch64 == *inductor-micro-benchmark* ]] 2025-09-07T07:35:39.4988361Z + [[ inductor_torchbench_perf_cpu_aarch64 == *huggingface* ]] 2025-09-07T07:35:39.4988727Z + [[ inductor_torchbench_perf_cpu_aarch64 == *timm* ]] 2025-09-07T07:35:39.4989080Z + [[ inductor_torchbench_perf_cpu_aarch64 == cachebench ]] 2025-09-07T07:35:39.4989468Z + [[ inductor_torchbench_perf_cpu_aarch64 == verify_cachebench ]] 2025-09-07T07:35:39.4989871Z + [[ inductor_torchbench_perf_cpu_aarch64 == *torchbench* ]] 2025-09-07T07:35:39.4990372Z + install_torchaudio 2025-09-07T07:35:39.4990569Z + local commit 2025-09-07T07:35:39.4990767Z ++ get_pinned_commit audio 2025-09-07T07:35:39.4991005Z ++ cat .github/ci_commit_pins/audio.txt 2025-09-07T07:35:39.5007023Z + commit=2e300559e4e123928a22187b8f59a5b56f57ddc8 2025-09-07T07:35:39.5007613Z + pip_build_and_install git+https://github.com/pytorch/audio.git@2e300559e4e123928a22187b8f59a5b56f57ddc8 dist/audio 2025-09-07T07:35:39.5008360Z + local build_target=git+https://github.com/pytorch/audio.git@2e300559e4e123928a22187b8f59a5b56f57ddc8 2025-09-07T07:35:39.5008833Z + local wheel_dir=dist/audio 2025-09-07T07:35:39.5009061Z + local found_whl=0 2025-09-07T07:35:39.5018593Z + for file in "${wheel_dir}"/*.whl 2025-09-07T07:35:39.5019090Z + [[ -f dist/audio/torchaudio-2.8.0a0+2e30055-cp310-cp310-linux_aarch64.whl ]] 2025-09-07T07:35:39.5019478Z + found_whl=1 2025-09-07T07:35:39.5019649Z + break 2025-09-07T07:35:39.5019807Z + '[' 1 == 0 ']' 2025-09-07T07:35:39.5020019Z + for file in "${wheel_dir}"/*.whl 2025-09-07T07:35:39.5020429Z + pip_install_whl dist/audio/torchaudio-2.8.0a0+2e30055-cp310-cp310-linux_aarch64.whl 2025-09-07T07:35:39.5020995Z + args=('dist/audio/torchaudio-2.8.0a0+2e30055-cp310-cp310-linux_aarch64.whl') 2025-09-07T07:35:39.5021380Z + local args 2025-09-07T07:35:39.5021707Z + [[ dist/audio/torchaudio-2.8.0a0+2e30055-cp310-cp310-linux_aarch64.whl == *\ * ]] 2025-09-07T07:35:39.5022121Z + for path in "${args[@]}" 2025-09-07T07:35:39.5022515Z + echo 'Installing dist/audio/torchaudio-2.8.0a0+2e30055-cp310-cp310-linux_aarch64.whl' 2025-09-07T07:35:39.5023088Z Installing dist/audio/torchaudio-2.8.0a0+2e30055-cp310-cp310-linux_aarch64.whl 2025-09-07T07:35:39.5023742Z + python3 -mpip install --no-index --no-deps dist/audio/torchaudio-2.8.0a0+2e30055-cp310-cp310-linux_aarch64.whl 2025-09-07T07:35:39.8713276Z Processing ./dist/audio/torchaudio-2.8.0a0+2e30055-cp310-cp310-linux_aarch64.whl 2025-09-07T07:35:39.8783426Z Installing collected packages: torchaudio 2025-09-07T07:35:40.0727098Z Successfully installed torchaudio-2.8.0a0+2e30055 2025-09-07T07:35:40.1104544Z + install_torchvision 2025-09-07T07:35:40.1104755Z + local orig_preload 2025-09-07T07:35:40.1104952Z + local commit 2025-09-07T07:35:40.1107763Z ++ get_pinned_commit vision 2025-09-07T07:35:40.1108020Z ++ cat .github/ci_commit_pins/vision.txt 2025-09-07T07:35:40.1117898Z + commit=966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-09-07T07:35:40.1118196Z + orig_preload= 2025-09-07T07:35:40.1118391Z + '[' -n '' ']' 2025-09-07T07:35:40.1118599Z + [[ linux-jammy-aarch64-py3.10 == *cuda* ]] 2025-09-07T07:35:40.1119178Z + pip_build_and_install git+https://github.com/pytorch/vision.git@966da7e46f65d6d49df3e31214470a4fe5cc8e66 dist/vision 2025-09-07T07:35:40.1119943Z + local build_target=git+https://github.com/pytorch/vision.git@966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-09-07T07:35:40.1120428Z + local wheel_dir=dist/vision 2025-09-07T07:35:40.1120654Z + local found_whl=0 2025-09-07T07:35:40.1120883Z + for file in "${wheel_dir}"/*.whl 2025-09-07T07:35:40.1121281Z + [[ -f dist/vision/torchvision-0.22.0a0+966da7e-cp310-cp310-linux_aarch64.whl ]] 2025-09-07T07:35:40.1121675Z + found_whl=1 2025-09-07T07:35:40.1121843Z + break 2025-09-07T07:35:40.1121994Z + '[' 1 == 0 ']' 2025-09-07T07:35:40.1122492Z + for file in "${wheel_dir}"/*.whl 2025-09-07T07:35:40.1122928Z + pip_install_whl dist/vision/torchvision-0.22.0a0+966da7e-cp310-cp310-linux_aarch64.whl 2025-09-07T07:35:40.1123515Z + args=('dist/vision/torchvision-0.22.0a0+966da7e-cp310-cp310-linux_aarch64.whl') 2025-09-07T07:35:40.1123905Z + local args 2025-09-07T07:35:40.1124247Z + [[ dist/vision/torchvision-0.22.0a0+966da7e-cp310-cp310-linux_aarch64.whl == *\ * ]] 2025-09-07T07:35:40.1124672Z + for path in "${args[@]}" 2025-09-07T07:35:40.1125079Z + echo 'Installing dist/vision/torchvision-0.22.0a0+966da7e-cp310-cp310-linux_aarch64.whl' 2025-09-07T07:35:40.1125680Z Installing dist/vision/torchvision-0.22.0a0+966da7e-cp310-cp310-linux_aarch64.whl 2025-09-07T07:35:40.1126551Z + python3 -mpip install --no-index --no-deps dist/vision/torchvision-0.22.0a0+966da7e-cp310-cp310-linux_aarch64.whl 2025-09-07T07:35:40.4811293Z Processing ./dist/vision/torchvision-0.22.0a0+966da7e-cp310-cp310-linux_aarch64.whl 2025-09-07T07:35:40.4909239Z Installing collected packages: torchvision 2025-09-07T07:35:41.0100594Z Successfully installed torchvision-0.22.0a0+966da7e 2025-09-07T07:35:41.0454191Z + '[' -n '' ']' 2025-09-07T07:35:41.0454377Z + id=7 2025-09-07T07:35:41.0454570Z + pip_install opencv-python==4.8.0.74 2025-09-07T07:35:41.0454925Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-09-07T07:35:41.0455362Z + python3 -m pip install --progress-bar off opencv-python==4.8.0.74 2025-09-07T07:35:41.5178085Z Collecting opencv-python==4.8.0.74 2025-09-07T07:35:41.5333279Z Downloading opencv_python-4.8.0.74-cp37-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl.metadata (19 kB) 2025-09-07T07:35:41.5427105Z Requirement already satisfied: numpy>=1.21.2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from opencv-python==4.8.0.74) (1.22.4) 2025-09-07T07:35:41.5504038Z Downloading opencv_python-4.8.0.74-cp37-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (41.0 MB) 2025-09-07T07:35:42.7989653Z Installing collected packages: opencv-python 2025-09-07T07:35:42.7991031Z Attempting uninstall: opencv-python 2025-09-07T07:35:42.8005857Z Found existing installation: opencv-python 4.11.0.86 2025-09-07T07:35:42.8088576Z Uninstalling opencv-python-4.11.0.86: 2025-09-07T07:35:42.8739337Z Successfully uninstalled opencv-python-4.11.0.86 2025-09-07T07:35:43.4160340Z Successfully installed opencv-python-4.8.0.74 2025-09-07T07:35:43.4948930Z + [[ inductor_torchbench_perf_cpu_aarch64 == *inductor_torchbench_smoketest_perf* ]] 2025-09-07T07:35:43.4949496Z + [[ inductor_torchbench_perf_cpu_aarch64 == *inductor_torchbench_cpu_smoketest_perf* ]] 2025-09-07T07:35:43.4950021Z + [[ inductor_torchbench_perf_cpu_aarch64 == *torchbench_gcp_smoketest* ]] 2025-09-07T07:35:43.4950464Z + [[ inductor_torchbench_perf_cpu_aarch64 != *cpu* ]] 2025-09-07T07:35:43.4950760Z + PYTHONPATH=/torchbench 2025-09-07T07:35:43.4950993Z + test_dynamo_benchmark torchbench 7 2025-09-07T07:35:43.4953533Z ++ pwd 2025-09-07T07:35:43.4954967Z + TEST_REPORTS_DIR=/var/lib/jenkins/workspace/test/test-reports 2025-09-07T07:35:43.4955302Z + local suite=torchbench 2025-09-07T07:35:43.4955514Z + shift 2025-09-07T07:35:43.4955676Z + local shard_id=7 2025-09-07T07:35:43.4955857Z + shift 2025-09-07T07:35:43.4956101Z + [[ inductor_torchbench_perf_cpu_aarch64 == *perf_compare* ]] 2025-09-07T07:35:43.4956476Z + [[ inductor_torchbench_perf_cpu_aarch64 == *perf* ]] 2025-09-07T07:35:43.4956821Z + [[ inductor_torchbench_perf_cpu_aarch64 == *b200* ]] 2025-09-07T07:35:43.4957171Z + test_single_dynamo_benchmark dashboard torchbench 7 2025-09-07T07:35:43.4958660Z ++ pwd 2025-09-07T07:35:43.4960378Z + TEST_REPORTS_DIR=/var/lib/jenkins/workspace/test/test-reports 2025-09-07T07:35:43.4960773Z + mkdir -p /var/lib/jenkins/workspace/test/test-reports 2025-09-07T07:35:43.4975524Z + local name=dashboard 2025-09-07T07:35:43.4975733Z + shift 2025-09-07T07:35:43.4975901Z + local suite=torchbench 2025-09-07T07:35:43.4976102Z + shift 2025-09-07T07:35:43.4976276Z + local shard_id=7 2025-09-07T07:35:43.4976730Z + shift 2025-09-07T07:35:43.4976930Z + partition_flags=() 2025-09-07T07:35:43.4977141Z + local partition_flags 2025-09-07T07:35:43.4977343Z + [[ -n 15 ]] 2025-09-07T07:35:43.4977515Z + [[ -n 7 ]] 2025-09-07T07:35:43.4977856Z + partition_flags=(--total-partitions "$NUM_TEST_SHARDS" --partition-id "$shard_id") 2025-09-07T07:35:43.4978354Z + [[ inductor_torchbench_perf_cpu_aarch64 == *perf_compare* ]] 2025-09-07T07:35:43.4978732Z + [[ inductor_torchbench_perf_cpu_aarch64 == *perf* ]] 2025-09-07T07:35:43.4979202Z + test_perf_for_dashboard torchbench --device cpu --total-partitions 15 --partition-id 7 2025-09-07T07:35:43.4979617Z ++ pwd 2025-09-07T07:35:43.4981053Z + TEST_REPORTS_DIR=/var/lib/jenkins/workspace/test/test-reports 2025-09-07T07:35:43.4981632Z + mkdir -p /var/lib/jenkins/workspace/test/test-reports 2025-09-07T07:35:43.4994292Z + local suite=torchbench 2025-09-07T07:35:43.4994532Z + shift 2025-09-07T07:35:43.4994705Z + local backend=inductor 2025-09-07T07:35:43.4994924Z + modes=() 2025-09-07T07:35:43.4995096Z + local modes 2025-09-07T07:35:43.4995656Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *training-true* ]] 2025-09-07T07:35:43.4996644Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *inference-true* ]] 2025-09-07T07:35:43.4997263Z + modes+=(inference) 2025-09-07T07:35:43.4997481Z + targets=('accuracy' 'performance') 2025-09-07T07:35:43.4997730Z + local targets 2025-09-07T07:35:43.4997917Z + local device=cuda 2025-09-07T07:35:43.4998163Z + [[ inductor_torchbench_perf_cpu_aarch64 == *cpu* ]] 2025-09-07T07:35:43.4998525Z + [[ inductor_torchbench_perf_cpu_aarch64 == *cpu_x86_zen* ]] 2025-09-07T07:35:43.4998908Z + [[ inductor_torchbench_perf_cpu_aarch64 == *cpu_x86* ]] 2025-09-07T07:35:43.4999280Z + [[ inductor_torchbench_perf_cpu_aarch64 == *cpu_aarch64* ]] 2025-09-07T07:35:43.4999595Z + device=cpu_aarch64 2025-09-07T07:35:43.4999805Z + test_inductor_set_cpu_affinity 2025-09-07T07:35:43.5000074Z ++ find /usr/lib -name libjemalloc.so.2 2025-09-07T07:35:43.5176126Z + JEMALLOC_LIB=/usr/lib/aarch64-linux-gnu/libjemalloc.so.2 2025-09-07T07:35:43.5176537Z + export LD_PRELOAD=/usr/lib/aarch64-linux-gnu/libjemalloc.so.2: 2025-09-07T07:35:43.5176944Z + LD_PRELOAD=/usr/lib/aarch64-linux-gnu/libjemalloc.so.2: 2025-09-07T07:35:43.5177528Z + export MALLOC_CONF=oversize_threshold:1,background_thread:true,metadata_thp:auto,dirty_decay_ms:-1,muzzy_decay_ms:-1 2025-09-07T07:35:43.5178300Z + MALLOC_CONF=oversize_threshold:1,background_thread:true,metadata_thp:auto,dirty_decay_ms:-1,muzzy_decay_ms:-1 2025-09-07T07:35:43.5178858Z + [[ inductor_torchbench_perf_cpu_aarch64 != *aarch64* ]] 2025-09-07T07:35:43.5179688Z ++ nproc 2025-09-07T07:35:43.5200369Z + cpus=64 2025-09-07T07:35:43.5205186Z ++ lscpu 2025-09-07T07:35:43.5206417Z ++ grep 'Thread(s) per core:' 2025-09-07T07:35:43.5207439Z ++ awk '{print $4}' 2025-09-07T07:35:43.5446861Z + thread_per_core=1 2025-09-07T07:35:43.5447082Z + cores=64 2025-09-07T07:35:43.5447330Z + [[ inductor_torchbench_perf_cpu_aarch64 == *aarch64* ]] 2025-09-07T07:35:43.5447635Z + [[ 64 -gt 16 ]] 2025-09-07T07:35:43.5447815Z + cores=16 2025-09-07T07:35:43.5447991Z + export OMP_NUM_THREADS=16 2025-09-07T07:35:43.5448212Z + OMP_NUM_THREADS=16 2025-09-07T07:35:43.5450624Z ++ python -c 'import os; print(min(os.sched_getaffinity(0)))' 2025-09-07T07:35:43.5762541Z + start_cpu=0 2025-09-07T07:35:43.5765472Z ++ python -c 'import os; print(max(os.sched_getaffinity(0)))' 2025-09-07T07:35:43.6073415Z + end_cpu=62 2025-09-07T07:35:43.6073636Z + export 'TASKSET=taskset -c 0-62' 2025-09-07T07:35:43.6073893Z + TASKSET='taskset -c 0-62' 2025-09-07T07:35:43.6074126Z + for mode in "${modes[@]}" 2025-09-07T07:35:43.6074355Z + [[ inference == \i\n\f\e\r\e\n\c\e ]] 2025-09-07T07:35:43.6074637Z + [[ cpu_aarch64 == \c\p\u\_\x\8\6 ]] 2025-09-07T07:35:43.6074881Z + dtype=bfloat16 2025-09-07T07:35:43.6075088Z + for target in "${targets[@]}" 2025-09-07T07:35:43.6075321Z + target_flag=('--accuracy') 2025-09-07T07:35:43.6075848Z + local target_flag 2025-09-07T07:35:43.6076084Z + [[ accuracy == \p\e\r\f\o\r\m\a\n\c\e ]] 2025-09-07T07:35:43.6076351Z + [[ accuracy == \a\c\c\u\r\a\c\y ]] 2025-09-07T07:35:43.6076627Z + target_flag+=(--no-translation-validation) 2025-09-07T07:35:43.6077284Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *freezing-true* ]] 2025-09-07T07:35:43.6078263Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *default-true* ]] 2025-09-07T07:35:43.6080002Z + taskset -c 0-62 python benchmarks/dynamo/torchbench.py --accuracy --no-translation-validation --inference --bfloat16 --backend inductor --disable-cudagraphs --device cpu --total-partitions 15 --partition-id 7 --output /var/lib/jenkins/workspace/test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_aarch64_accuracy.csv 2025-09-07T07:35:43.9377764Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:35:43.9378939Z import pynvml # type: ignore[import] 2025-09-07T07:35:46.9095472Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:35:46.9096661Z import pynvml # type: ignore[import] 2025-09-07T07:35:49.4996194Z 2025-09-07T07:35:51.3507376Z loading model: 0it [00:00, ?it/s]Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T07:35:51.3509414Z WARNING:transformers.models.reformer.modeling_reformer:Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T07:35:51.5304369Z 2025-09-07T07:35:51.5304553Z loading model: 0it [00:02, ?it/s] 2025-09-07T07:35:51.5343583Z cpu eval hf_Reformer 2025-09-07T07:35:53.4916834Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:35:53.8156782Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:35:54.1365861Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:36:02.0128802Z W0907 07:36:02.012000 512 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] Graph break from `Tensor.item()`, consider setting: 2025-09-07T07:36:02.0129771Z W0907 07:36:02.012000 512 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] torch._dynamo.config.capture_scalar_outputs = True 2025-09-07T07:36:02.0130498Z W0907 07:36:02.012000 512 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] or: 2025-09-07T07:36:02.0131193Z W0907 07:36:02.012000 512 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1 2025-09-07T07:36:02.0132038Z W0907 07:36:02.012000 512 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] to include these operations in the captured graph. 2025-09-07T07:36:02.0132734Z W0907 07:36:02.012000 512 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T07:36:02.0133388Z W0907 07:36:02.012000 512 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] Graph break: from user code at: 2025-09-07T07:36:02.0136163Z W0907 07:36:02.012000 512 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 2132, in torch_dynamo_resume_in_forward_at_2104 2025-09-07T07:36:02.0137505Z W0907 07:36:02.012000 512 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] embedding_output = self.embeddings( 2025-09-07T07:36:02.0138654Z W0907 07:36:02.012000 512 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 369, in forward 2025-09-07T07:36:02.0139871Z W0907 07:36:02.012000 512 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] position_embeddings = self.position_embeddings(position_ids) 2025-09-07T07:36:02.0141087Z W0907 07:36:02.012000 512 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 293, in forward 2025-09-07T07:36:02.0142442Z W0907 07:36:02.012000 512 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] max_position_id = position_ids.max().item() 2025-09-07T07:36:02.0143120Z W0907 07:36:02.012000 512 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T07:36:02.0143652Z W0907 07:36:02.012000 512 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T07:36:15.4104189Z pass 2025-09-07T07:36:15.4104676Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:36:16.9097192Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:36:16.9098429Z import pynvml # type: ignore[import] 2025-09-07T07:36:19.4719640Z 2025-09-07T07:36:20.5415826Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:36:20.5416072Z 2025-09-07T07:36:20.5417607Z config.json: 0% 0.00/615 [00:00bmhtr", vectors, random_rotations) 2025-09-07T07:40:01.0570379Z 2025-09-07T07:40:01.0570518Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0571283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0572032Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0572664Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0573272Z layer_outputs = layer( 2025-09-07T07:40:01.0573840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0574444Z attn_outputs = self.attention( 2025-09-07T07:40:01.0575034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0575668Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0576298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T07:40:01.0577013Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T07:40:01.0577749Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T07:40:01.0578492Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T07:40:01.0578789Z 2025-09-07T07:40:01.0578926Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0579695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0580445Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0581071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0581669Z layer_outputs = layer( 2025-09-07T07:40:01.0582236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0582844Z attn_outputs = self.attention( 2025-09-07T07:40:01.0583432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0584070Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0584898Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T07:40:01.0585639Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T07:40:01.0586382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T07:40:01.0587035Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T07:40:01.0587254Z 2025-09-07T07:40:01.0587388Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0588152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0588992Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0589624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0590221Z layer_outputs = layer( 2025-09-07T07:40:01.0590793Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0591403Z attn_outputs = self.attention( 2025-09-07T07:40:01.0591995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0592629Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0593254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T07:40:01.0594046Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T07:40:01.0594986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T07:40:01.0595784Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T07:40:01.0596464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T07:40:01.0597226Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T07:40:01.0597541Z 2025-09-07T07:40:01.0597640Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0597960Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0598727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0599478Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0600108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0600707Z layer_outputs = layer( 2025-09-07T07:40:01.0601279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0601891Z attn_outputs = self.attention( 2025-09-07T07:40:01.0602484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0603115Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0603743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T07:40:01.0604530Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T07:40:01.0605469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T07:40:01.0606261Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T07:40:01.0607044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T07:40:01.0607728Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T07:40:01.0607941Z 2025-09-07T07:40:01.0608040Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0608321Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0608638Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0609405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0610262Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0610898Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0611503Z layer_outputs = layer( 2025-09-07T07:40:01.0612067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0612678Z attn_outputs = self.attention( 2025-09-07T07:40:01.0613274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0613914Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0614544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.0615189Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.0615833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T07:40:01.0616657Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.0617511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T07:40:01.0618264Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T07:40:01.0618573Z 2025-09-07T07:40:01.0618671Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0618951Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0619265Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0620038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0620788Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0621415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0622018Z layer_outputs = layer( 2025-09-07T07:40:01.0622590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0623203Z attn_outputs = self.attention( 2025-09-07T07:40:01.0623797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0624429Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0625057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.0625707Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.0626354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T07:40:01.0627179Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.0628141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T07:40:01.0628788Z return torch.cat(slices, dim=3) 2025-09-07T07:40:01.0628967Z 2025-09-07T07:40:01.0629064Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0629383Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0630147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0630895Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0631526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0632223Z layer_outputs = layer( 2025-09-07T07:40:01.0632794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0633399Z attn_outputs = self.attention( 2025-09-07T07:40:01.0633993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0634629Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0635256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 541, in forward 2025-09-07T07:40:01.0635876Z value_vectors = self.value(hidden_states) 2025-09-07T07:40:01.0636074Z 2025-09-07T07:40:01.0636206Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0636971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0637727Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0638364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0638974Z layer_outputs = layer( 2025-09-07T07:40:01.0639541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0640151Z attn_outputs = self.attention( 2025-09-07T07:40:01.0640748Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0641385Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0642018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.0642668Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.0643317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T07:40:01.0644093Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.0644899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T07:40:01.0645652Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T07:40:01.0645957Z 2025-09-07T07:40:01.0646053Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0646367Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0647136Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0647887Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0648663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0649272Z layer_outputs = layer( 2025-09-07T07:40:01.0649840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0650448Z attn_outputs = self.attention( 2025-09-07T07:40:01.0651042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0651676Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0652299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.0653031Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.0653674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T07:40:01.0654451Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.0655250Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T07:40:01.0655879Z return torch.cat(slices, dim=3) 2025-09-07T07:40:01.0656056Z 2025-09-07T07:40:01.0656191Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0656954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0657698Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0658332Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0658937Z layer_outputs = layer( 2025-09-07T07:40:01.0659507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0660117Z attn_outputs = self.attention( 2025-09-07T07:40:01.0660709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0661346Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0661970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T07:40:01.0662783Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T07:40:01.0663657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T07:40:01.0664358Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T07:40:01.0664572Z 2025-09-07T07:40:01.0664710Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0665569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0666324Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0666956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0667557Z layer_outputs = layer( 2025-09-07T07:40:01.0668124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0668733Z attn_outputs = self.attention( 2025-09-07T07:40:01.0669334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0669971Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0670764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.0671437Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.0672085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T07:40:01.0672812Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T07:40:01.0673120Z 2025-09-07T07:40:01.0673255Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0674017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0674881Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0675513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0676117Z layer_outputs = layer( 2025-09-07T07:40:01.0676682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0677292Z attn_outputs = self.attention( 2025-09-07T07:40:01.0677884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0678522Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0679153Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.0679803Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.0680449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T07:40:01.0681241Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.0682065Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T07:40:01.0682817Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T07:40:01.0683121Z 2025-09-07T07:40:01.0683223Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0683538Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0684298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0685045Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0685679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0686281Z layer_outputs = layer( 2025-09-07T07:40:01.0686850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0687457Z attn_outputs = self.attention( 2025-09-07T07:40:01.0688049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0688687Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0689321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.0689971Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.0690611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T07:40:01.0691283Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T07:40:01.0691535Z 2025-09-07T07:40:01.0691802Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0692575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0693322Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0693947Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0694550Z layer_outputs = layer( 2025-09-07T07:40:01.0695116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0695727Z attn_outputs = self.attention( 2025-09-07T07:40:01.0696405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0697039Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0697670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.0698318Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.0698960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T07:40:01.0699621Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T07:40:01.0699872Z 2025-09-07T07:40:01.0699969Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0700285Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0701051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0701802Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0702430Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0703036Z layer_outputs = layer( 2025-09-07T07:40:01.0703607Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0704222Z attn_outputs = self.attention( 2025-09-07T07:40:01.0704819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0705450Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0706078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T07:40:01.0706904Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T07:40:01.0707795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T07:40:01.0708577Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T07:40:01.0708869Z 2025-09-07T07:40:01.0709003Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0709770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0710522Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0711155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0711759Z layer_outputs = layer( 2025-09-07T07:40:01.0712330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0712939Z attn_outputs = self.attention( 2025-09-07T07:40:01.0713641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T07:40:01.0714364Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T07:40:01.0715078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T07:40:01.0715701Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.0715901Z 2025-09-07T07:40:01.0716035Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0716801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0717628Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0718261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0718867Z layer_outputs = layer( 2025-09-07T07:40:01.0719443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.0720129Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.0720815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T07:40:01.0721432Z return apply_chunking_to_forward( 2025-09-07T07:40:01.0722011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:01.0722604Z return forward_fn(*input_tensors) 2025-09-07T07:40:01.0723236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T07:40:01.0723888Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.0724506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T07:40:01.0725129Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.0725330Z 2025-09-07T07:40:01.0725463Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0726230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0726975Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0727602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0728205Z layer_outputs = layer( 2025-09-07T07:40:01.0728772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.0729460Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.0730143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T07:40:01.0730751Z return apply_chunking_to_forward( 2025-09-07T07:40:01.0731334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:01.0731929Z return forward_fn(*input_tensors) 2025-09-07T07:40:01.0732556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T07:40:01.0733206Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.0733825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T07:40:01.0734450Z hidden_states = self.act_fn(hidden_states) 2025-09-07T07:40:01.0734765Z 2025-09-07T07:40:01.0734906Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0735672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0736413Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0737039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0737638Z layer_outputs = layer( 2025-09-07T07:40:01.0738205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.0738980Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.0739661Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T07:40:01.0740277Z return apply_chunking_to_forward( 2025-09-07T07:40:01.0740857Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:01.0741445Z return forward_fn(*input_tensors) 2025-09-07T07:40:01.0742066Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T07:40:01.0742707Z return self.output(hidden_states) 2025-09-07T07:40:01.0743307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T07:40:01.0743926Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.0744130Z 2025-09-07T07:40:01.0744263Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0745032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0745774Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0746402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0747003Z layer_outputs = layer( 2025-09-07T07:40:01.0747571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0748181Z attn_outputs = self.attention( 2025-09-07T07:40:01.0748773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0749414Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0750049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1198, in forward 2025-09-07T07:40:01.0750675Z query_vectors = self.query(hidden_states) 2025-09-07T07:40:01.0750871Z 2025-09-07T07:40:01.0751010Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0751769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0752516Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0753147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0753748Z layer_outputs = layer( 2025-09-07T07:40:01.0754313Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0754922Z attn_outputs = self.attention( 2025-09-07T07:40:01.0755512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0756266Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0756917Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1199, in forward 2025-09-07T07:40:01.0757529Z key_vectors = self.key(hidden_states) 2025-09-07T07:40:01.0757718Z 2025-09-07T07:40:01.0757816Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0758096Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0758411Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0759171Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0760001Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0760630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0761235Z layer_outputs = layer( 2025-09-07T07:40:01.0761799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0762409Z attn_outputs = self.attention( 2025-09-07T07:40:01.0763001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0763638Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0764267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T07:40:01.0765041Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.0765938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T07:40:01.0766703Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T07:40:01.0767011Z 2025-09-07T07:40:01.0767109Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0767385Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0767695Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0768461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0769208Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0769845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0770451Z layer_outputs = layer( 2025-09-07T07:40:01.0771012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0771625Z attn_outputs = self.attention( 2025-09-07T07:40:01.0772220Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0772858Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0773490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T07:40:01.0774261Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.0775068Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T07:40:01.0775705Z return torch.cat(slices, dim=3) 2025-09-07T07:40:01.0775875Z 2025-09-07T07:40:01.0776011Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0776960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0777718Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0778350Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0778951Z layer_outputs = layer( 2025-09-07T07:40:01.0779515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0780124Z attn_outputs = self.attention( 2025-09-07T07:40:01.0780713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0781494Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0782128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T07:40:01.0782857Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T07:40:01.0783164Z 2025-09-07T07:40:01.0783297Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0784061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0784810Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0785439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0786044Z layer_outputs = layer( 2025-09-07T07:40:01.0786611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0787223Z attn_outputs = self.attention( 2025-09-07T07:40:01.0787818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0788457Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0789089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T07:40:01.0789808Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T07:40:01.0790113Z 2025-09-07T07:40:01.0790249Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0791011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0791766Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0792393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0792989Z layer_outputs = layer( 2025-09-07T07:40:01.0793555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0794165Z attn_outputs = self.attention( 2025-09-07T07:40:01.0794759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0795394Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0796024Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1200, in forward 2025-09-07T07:40:01.0796648Z value_vectors = self.value(hidden_states) 2025-09-07T07:40:01.0796847Z 2025-09-07T07:40:01.0796949Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0797227Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0797539Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0798413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0799175Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0799804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0800404Z layer_outputs = layer( 2025-09-07T07:40:01.0800967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0801576Z attn_outputs = self.attention( 2025-09-07T07:40:01.0802170Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0802887Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0803522Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T07:40:01.0804318Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.0805137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T07:40:01.0805887Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T07:40:01.0806193Z 2025-09-07T07:40:01.0806294Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0806571Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0806880Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0807643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0808394Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0809031Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0809636Z layer_outputs = layer( 2025-09-07T07:40:01.0810201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0810812Z attn_outputs = self.attention( 2025-09-07T07:40:01.0811406Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0812042Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0812672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T07:40:01.0813347Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T07:40:01.0813600Z 2025-09-07T07:40:01.0813734Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0814503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0815252Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0815883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0816484Z layer_outputs = layer( 2025-09-07T07:40:01.0817049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0817660Z attn_outputs = self.attention( 2025-09-07T07:40:01.0818252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0818889Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0819628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T07:40:01.0820309Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T07:40:01.0820557Z 2025-09-07T07:40:01.0820693Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0821456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0822196Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0822828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0823526Z layer_outputs = layer( 2025-09-07T07:40:01.0824093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0824702Z attn_outputs = self.attention( 2025-09-07T07:40:01.0825296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0825930Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0826567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T07:40:01.0827386Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T07:40:01.0828270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T07:40:01.0829048Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T07:40:01.0829347Z 2025-09-07T07:40:01.0829482Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0830252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0830995Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0831626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0832224Z layer_outputs = layer( 2025-09-07T07:40:01.0832791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0833401Z attn_outputs = self.attention( 2025-09-07T07:40:01.0833992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T07:40:01.0834707Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T07:40:01.0835418Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T07:40:01.0836047Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.0836247Z 2025-09-07T07:40:01.0836381Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0837143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0837887Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0838513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0839112Z layer_outputs = layer( 2025-09-07T07:40:01.0839680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.0840370Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.0841160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T07:40:01.0841786Z return apply_chunking_to_forward( 2025-09-07T07:40:01.0842368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:01.0842959Z return forward_fn(*input_tensors) 2025-09-07T07:40:01.0843581Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T07:40:01.0844234Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.0844852Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T07:40:01.0845562Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.0845761Z 2025-09-07T07:40:01.0845896Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0846662Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0847407Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0848035Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0848635Z layer_outputs = layer( 2025-09-07T07:40:01.0849201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.0849883Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.0850565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T07:40:01.0851178Z return apply_chunking_to_forward( 2025-09-07T07:40:01.0851763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:01.0852354Z return forward_fn(*input_tensors) 2025-09-07T07:40:01.0852977Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T07:40:01.0853623Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.0854242Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T07:40:01.0854867Z hidden_states = self.act_fn(hidden_states) 2025-09-07T07:40:01.0855064Z 2025-09-07T07:40:01.0855202Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0855968Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0856714Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0857347Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0857948Z layer_outputs = layer( 2025-09-07T07:40:01.0858516Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.0859199Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.0859884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T07:40:01.0860502Z return apply_chunking_to_forward( 2025-09-07T07:40:01.0861086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:01.0861677Z return forward_fn(*input_tensors) 2025-09-07T07:40:01.0862407Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T07:40:01.0863053Z return self.output(hidden_states) 2025-09-07T07:40:01.0863658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T07:40:01.0864281Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.0864476Z 2025-09-07T07:40:01.0864611Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0865444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0866191Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0866978Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0867579Z layer_outputs = layer( 2025-09-07T07:40:01.0868149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0868756Z attn_outputs = self.attention( 2025-09-07T07:40:01.0869348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0869983Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0870620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 540, in forward 2025-09-07T07:40:01.0871262Z query_key_vectors = self.query_key(hidden_states) 2025-09-07T07:40:01.0871479Z 2025-09-07T07:40:01.0871617Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0872385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0873138Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0873771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0874374Z layer_outputs = layer( 2025-09-07T07:40:01.0874942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0875554Z attn_outputs = self.attention( 2025-09-07T07:40:01.0876148Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0876789Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0877427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T07:40:01.0878141Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T07:40:01.0878890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 749, in _hash_vectors 2025-09-07T07:40:01.0879702Z random_rotations = torch.randn(rotations_shape, device=vectors.device, dtype=vectors.dtype) 2025-09-07T07:40:01.0880070Z 2025-09-07T07:40:01.0880207Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0880973Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0881720Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0882353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0882960Z layer_outputs = layer( 2025-09-07T07:40:01.0883688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0884319Z attn_outputs = self.attention( 2025-09-07T07:40:01.0884910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0885543Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0886172Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T07:40:01.0886889Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T07:40:01.0887630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T07:40:01.0888476Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T07:40:01.0888795Z 2025-09-07T07:40:01.0888930Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0889700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0890450Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0891083Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0891679Z layer_outputs = layer( 2025-09-07T07:40:01.0892244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0892853Z attn_outputs = self.attention( 2025-09-07T07:40:01.0893443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0894083Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0894710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T07:40:01.0895427Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T07:40:01.0896171Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T07:40:01.0896908Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T07:40:01.0897202Z 2025-09-07T07:40:01.0897339Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0898103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0898853Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0899487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0900091Z layer_outputs = layer( 2025-09-07T07:40:01.0900660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0901267Z attn_outputs = self.attention( 2025-09-07T07:40:01.0901863Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0902502Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0903134Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T07:40:01.0903850Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T07:40:01.0904597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T07:40:01.0905370Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T07:40:01.0905596Z 2025-09-07T07:40:01.0905731Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0906494Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0907244Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0907876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0908478Z layer_outputs = layer( 2025-09-07T07:40:01.0909045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0909797Z attn_outputs = self.attention( 2025-09-07T07:40:01.0910391Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0911032Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0911663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T07:40:01.0912453Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T07:40:01.0913389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T07:40:01.0914182Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T07:40:01.0914858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T07:40:01.0915625Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T07:40:01.0915941Z 2025-09-07T07:40:01.0916036Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0916357Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0917123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0917870Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0918506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0919107Z layer_outputs = layer( 2025-09-07T07:40:01.0919672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0920285Z attn_outputs = self.attention( 2025-09-07T07:40:01.0920876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0921515Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0922152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T07:40:01.0922939Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T07:40:01.0923872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T07:40:01.0924669Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T07:40:01.0925346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T07:40:01.0926017Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T07:40:01.0926227Z 2025-09-07T07:40:01.0926327Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0926601Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0927021Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0927797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0928544Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0929175Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0929775Z layer_outputs = layer( 2025-09-07T07:40:01.0930343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0931034Z attn_outputs = self.attention( 2025-09-07T07:40:01.0931625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0932259Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0932885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.0933536Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.0934179Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T07:40:01.0934993Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.0935839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T07:40:01.0936601Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T07:40:01.0936909Z 2025-09-07T07:40:01.0937006Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0937288Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0937602Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0938364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0939112Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0939749Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0940350Z layer_outputs = layer( 2025-09-07T07:40:01.0940916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0941530Z attn_outputs = self.attention( 2025-09-07T07:40:01.0942124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0942765Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0943393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.0944040Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.0944679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T07:40:01.0945495Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.0946341Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T07:40:01.0946981Z return torch.cat(slices, dim=3) 2025-09-07T07:40:01.0947154Z 2025-09-07T07:40:01.0947254Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0947565Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0948443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0949199Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0949828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0950430Z layer_outputs = layer( 2025-09-07T07:40:01.0950993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0951605Z attn_outputs = self.attention( 2025-09-07T07:40:01.0952200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0952921Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0953549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 541, in forward 2025-09-07T07:40:01.0954170Z value_vectors = self.value(hidden_states) 2025-09-07T07:40:01.0954370Z 2025-09-07T07:40:01.0954504Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0955270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0956019Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0956644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0957248Z layer_outputs = layer( 2025-09-07T07:40:01.0957815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0958424Z attn_outputs = self.attention( 2025-09-07T07:40:01.0959019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0959653Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0960283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.0960934Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.0961577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T07:40:01.0962353Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.0963157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T07:40:01.0963915Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T07:40:01.0964227Z 2025-09-07T07:40:01.0964324Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.0964640Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0965486Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0966237Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0966871Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0967473Z layer_outputs = layer( 2025-09-07T07:40:01.0968038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0968656Z attn_outputs = self.attention( 2025-09-07T07:40:01.0969412Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0970059Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0970687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.0971330Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.0971974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T07:40:01.0972746Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.0973554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T07:40:01.0974302Z return torch.cat(slices, dim=3) 2025-09-07T07:40:01.0974475Z 2025-09-07T07:40:01.0974611Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0975380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0976128Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0976761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0977364Z layer_outputs = layer( 2025-09-07T07:40:01.0977931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0978537Z attn_outputs = self.attention( 2025-09-07T07:40:01.0979133Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0979769Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0980399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T07:40:01.0981213Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T07:40:01.0982080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T07:40:01.0982775Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T07:40:01.0982990Z 2025-09-07T07:40:01.0983126Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0983890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0984644Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0985277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0985880Z layer_outputs = layer( 2025-09-07T07:40:01.0986446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0987059Z attn_outputs = self.attention( 2025-09-07T07:40:01.0987655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0988302Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0988933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.0989584Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.0990232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T07:40:01.0991088Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T07:40:01.0991400Z 2025-09-07T07:40:01.0991537Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.0992299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.0993051Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.0993683Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.0994286Z layer_outputs = layer( 2025-09-07T07:40:01.0994848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.0995538Z attn_outputs = self.attention( 2025-09-07T07:40:01.0996131Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.0996773Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.0997403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.0998048Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.0998690Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T07:40:01.0999485Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.1000309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T07:40:01.1001063Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T07:40:01.1001369Z 2025-09-07T07:40:01.1001467Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1001609Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1002145Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1002264Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1002688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1002767Z layer_outputs = layer( 2025-09-07T07:40:01.1003192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1012863Z attn_outputs = self.attention( 2025-09-07T07:40:01.1013417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1013558Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1013996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.1014134Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.1014554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T07:40:01.1014709Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T07:40:01.1014714Z 2025-09-07T07:40:01.1014851Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1015395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1015515Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1016116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1016208Z layer_outputs = layer( 2025-09-07T07:40:01.1016635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1016727Z attn_outputs = self.attention( 2025-09-07T07:40:01.1017146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1017266Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1017686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.1017900Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.1018322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T07:40:01.1018478Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T07:40:01.1018482Z 2025-09-07T07:40:01.1018580Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1018713Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1019251Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1019362Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1019787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1019865Z layer_outputs = layer( 2025-09-07T07:40:01.1020290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1020380Z attn_outputs = self.attention( 2025-09-07T07:40:01.1020805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1020923Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1021339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T07:40:01.1021639Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T07:40:01.1022127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T07:40:01.1022322Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T07:40:01.1022331Z 2025-09-07T07:40:01.1022468Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1023007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1023123Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1023551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1023631Z layer_outputs = layer( 2025-09-07T07:40:01.1024050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1024137Z attn_outputs = self.attention( 2025-09-07T07:40:01.1024561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T07:40:01.1024756Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T07:40:01.1025181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T07:40:01.1025368Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.1025373Z 2025-09-07T07:40:01.1025512Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1026048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1026161Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1026586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1026664Z layer_outputs = layer( 2025-09-07T07:40:01.1027090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1601, in forward 2025-09-07T07:40:01.1027271Z attn_output = prev_attn_output + attn_output 2025-09-07T07:40:01.1027276Z 2025-09-07T07:40:01.1027410Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1027955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1028066Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1028491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1028567Z layer_outputs = layer( 2025-09-07T07:40:01.1028992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.1029156Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.1029579Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T07:40:01.1029676Z return apply_chunking_to_forward( 2025-09-07T07:40:01.1030083Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:01.1030179Z return forward_fn(*input_tensors) 2025-09-07T07:40:01.1030628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T07:40:01.1030729Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.1031152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T07:40:01.1031253Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.1031258Z 2025-09-07T07:40:01.1031392Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1031932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1032052Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1032472Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1032550Z layer_outputs = layer( 2025-09-07T07:40:01.1032976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.1033140Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.1033564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T07:40:01.1033655Z return apply_chunking_to_forward( 2025-09-07T07:40:01.1034067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:01.1034159Z return forward_fn(*input_tensors) 2025-09-07T07:40:01.1034698Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T07:40:01.1034813Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.1035237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T07:40:01.1035344Z hidden_states = self.act_fn(hidden_states) 2025-09-07T07:40:01.1035349Z 2025-09-07T07:40:01.1035483Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1036018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1036204Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1036624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1036704Z layer_outputs = layer( 2025-09-07T07:40:01.1037127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.1037292Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.1037712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T07:40:01.1037802Z return apply_chunking_to_forward( 2025-09-07T07:40:01.1038206Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:01.1038297Z return forward_fn(*input_tensors) 2025-09-07T07:40:01.1038752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T07:40:01.1038844Z return self.output(hidden_states) 2025-09-07T07:40:01.1039267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T07:40:01.1039372Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.1039377Z 2025-09-07T07:40:01.1039509Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1040048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1040160Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1040583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1040666Z layer_outputs = layer( 2025-09-07T07:40:01.1041086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.1041254Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.1041259Z 2025-09-07T07:40:01.1041392Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1041931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1042042Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1042466Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1042544Z layer_outputs = layer( 2025-09-07T07:40:01.1042964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1043060Z attn_outputs = self.attention( 2025-09-07T07:40:01.1043479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1043687Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1044114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1198, in forward 2025-09-07T07:40:01.1044218Z query_vectors = self.query(hidden_states) 2025-09-07T07:40:01.1044228Z 2025-09-07T07:40:01.1044361Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1044894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1045009Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1045429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1045575Z layer_outputs = layer( 2025-09-07T07:40:01.1045998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1046086Z attn_outputs = self.attention( 2025-09-07T07:40:01.1046508Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1046622Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1047045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1199, in forward 2025-09-07T07:40:01.1047138Z key_vectors = self.key(hidden_states) 2025-09-07T07:40:01.1047143Z 2025-09-07T07:40:01.1047240Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1047335Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1047473Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1048016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1048126Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1048551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1048630Z layer_outputs = layer( 2025-09-07T07:40:01.1049049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1049141Z attn_outputs = self.attention( 2025-09-07T07:40:01.1049565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1049683Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1050109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T07:40:01.1050373Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.1050824Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T07:40:01.1051026Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T07:40:01.1051032Z 2025-09-07T07:40:01.1051130Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1051224Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1051361Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1051899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1052015Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1052440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1052611Z layer_outputs = layer( 2025-09-07T07:40:01.1053044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1053133Z attn_outputs = self.attention( 2025-09-07T07:40:01.1053560Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1053674Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1054092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T07:40:01.1054355Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.1054875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T07:40:01.1054968Z return torch.cat(slices, dim=3) 2025-09-07T07:40:01.1054974Z 2025-09-07T07:40:01.1055108Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1055648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1055760Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1056181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1056262Z layer_outputs = layer( 2025-09-07T07:40:01.1056683Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1056779Z attn_outputs = self.attention( 2025-09-07T07:40:01.1057198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1057315Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1057740Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T07:40:01.1057949Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T07:40:01.1057954Z 2025-09-07T07:40:01.1058090Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1058627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1058742Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1059168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1059246Z layer_outputs = layer( 2025-09-07T07:40:01.1059673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1059761Z attn_outputs = self.attention( 2025-09-07T07:40:01.1060185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1060300Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1060721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T07:40:01.1060927Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T07:40:01.1060933Z 2025-09-07T07:40:01.1061063Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1061606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1061791Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1062237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1062317Z layer_outputs = layer( 2025-09-07T07:40:01.1062740Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1062832Z attn_outputs = self.attention( 2025-09-07T07:40:01.1063256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1063373Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1063867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1200, in forward 2025-09-07T07:40:01.1063971Z value_vectors = self.value(hidden_states) 2025-09-07T07:40:01.1063976Z 2025-09-07T07:40:01.1064076Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1064172Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1064309Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1064846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1064962Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1065505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1065583Z layer_outputs = layer( 2025-09-07T07:40:01.1066008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1066100Z attn_outputs = self.attention( 2025-09-07T07:40:01.1066526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1066641Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1067064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T07:40:01.1067342Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.1067792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T07:40:01.1067999Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T07:40:01.1068009Z 2025-09-07T07:40:01.1068106Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1068203Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1068335Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1068877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1068991Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1069414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1069495Z layer_outputs = layer( 2025-09-07T07:40:01.1069917Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1070007Z attn_outputs = self.attention( 2025-09-07T07:40:01.1070426Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1070543Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1071124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T07:40:01.1071292Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T07:40:01.1071297Z 2025-09-07T07:40:01.1071434Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1071968Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1072082Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1072505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1072584Z layer_outputs = layer( 2025-09-07T07:40:01.1073125Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1073213Z attn_outputs = self.attention( 2025-09-07T07:40:01.1073643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1073758Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1074178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T07:40:01.1074329Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T07:40:01.1074334Z 2025-09-07T07:40:01.1074465Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1075004Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1075120Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1075544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1075626Z layer_outputs = layer( 2025-09-07T07:40:01.1076046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1076136Z attn_outputs = self.attention( 2025-09-07T07:40:01.1076556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1076672Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1077091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T07:40:01.1077395Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T07:40:01.1077884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T07:40:01.1078083Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T07:40:01.1078088Z 2025-09-07T07:40:01.1078225Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1078758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1078875Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1079296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1079378Z layer_outputs = layer( 2025-09-07T07:40:01.1079797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1079889Z attn_outputs = self.attention( 2025-09-07T07:40:01.1080392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T07:40:01.1080592Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T07:40:01.1081019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T07:40:01.1081124Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.1081128Z 2025-09-07T07:40:01.1081261Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1081801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1081914Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1082449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1082529Z layer_outputs = layer( 2025-09-07T07:40:01.1082958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.1083121Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.1083545Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T07:40:01.1083641Z return apply_chunking_to_forward( 2025-09-07T07:40:01.1084044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:01.1084140Z return forward_fn(*input_tensors) 2025-09-07T07:40:01.1084589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T07:40:01.1084696Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.1085124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T07:40:01.1085228Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.1085233Z 2025-09-07T07:40:01.1085369Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1085902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1086017Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1086443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1086521Z layer_outputs = layer( 2025-09-07T07:40:01.1086950Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.1087113Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.1087540Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T07:40:01.1087634Z return apply_chunking_to_forward( 2025-09-07T07:40:01.1088039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:01.1088131Z return forward_fn(*input_tensors) 2025-09-07T07:40:01.1088577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T07:40:01.1088683Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.1089104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T07:40:01.1089217Z hidden_states = self.act_fn(hidden_states) 2025-09-07T07:40:01.1089222Z 2025-09-07T07:40:01.1089356Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1089984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1090126Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1090551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1090631Z layer_outputs = layer( 2025-09-07T07:40:01.1091051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.1091217Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.1091712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T07:40:01.1091804Z return apply_chunking_to_forward( 2025-09-07T07:40:01.1092216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:01.1092308Z return forward_fn(*input_tensors) 2025-09-07T07:40:01.1092756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T07:40:01.1092849Z return self.output(hidden_states) 2025-09-07T07:40:01.1093268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T07:40:01.1093372Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.1093377Z 2025-09-07T07:40:01.1093509Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1094051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1094168Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1094591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1094668Z layer_outputs = layer( 2025-09-07T07:40:01.1095088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1095180Z attn_outputs = self.attention( 2025-09-07T07:40:01.1095602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1095720Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1096141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 540, in forward 2025-09-07T07:40:01.1096265Z query_key_vectors = self.query_key(hidden_states) 2025-09-07T07:40:01.1096269Z 2025-09-07T07:40:01.1096404Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1096938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1097053Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1097473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1097556Z layer_outputs = layer( 2025-09-07T07:40:01.1097977Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1098068Z attn_outputs = self.attention( 2025-09-07T07:40:01.1098494Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1098608Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1099109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T07:40:01.1099320Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T07:40:01.1099767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 749, in _hash_vectors 2025-09-07T07:40:01.1100035Z random_rotations = torch.randn(rotations_shape, device=vectors.device, dtype=vectors.dtype) 2025-09-07T07:40:01.1100040Z 2025-09-07T07:40:01.1100173Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1100713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1100895Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1101327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1101407Z layer_outputs = layer( 2025-09-07T07:40:01.1101832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1101920Z attn_outputs = self.attention( 2025-09-07T07:40:01.1102339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1102457Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1102877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T07:40:01.1103086Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T07:40:01.1103532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T07:40:01.1103746Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T07:40:01.1103752Z 2025-09-07T07:40:01.1103884Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1104417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1104534Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1104957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1105037Z layer_outputs = layer( 2025-09-07T07:40:01.1105463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1105551Z attn_outputs = self.attention( 2025-09-07T07:40:01.1105979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1106095Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1106516Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T07:40:01.1106715Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T07:40:01.1107160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T07:40:01.1107355Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T07:40:01.1107364Z 2025-09-07T07:40:01.1107497Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1108121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1108241Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1108671Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1108749Z layer_outputs = layer( 2025-09-07T07:40:01.1109172Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1109260Z attn_outputs = self.attention( 2025-09-07T07:40:01.1109682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1109875Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1110293Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T07:40:01.1110500Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T07:40:01.1110945Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T07:40:01.1111063Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T07:40:01.1111071Z 2025-09-07T07:40:01.1111204Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1111740Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1111858Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1112278Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1112364Z layer_outputs = layer( 2025-09-07T07:40:01.1112792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1112879Z attn_outputs = self.attention( 2025-09-07T07:40:01.1113302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1113417Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1113840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T07:40:01.1114112Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T07:40:01.1114682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T07:40:01.1114819Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T07:40:01.1115275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T07:40:01.1115491Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T07:40:01.1115496Z 2025-09-07T07:40:01.1115592Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1115728Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1116264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1116380Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1116804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1116885Z layer_outputs = layer( 2025-09-07T07:40:01.1117311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1117489Z attn_outputs = self.attention( 2025-09-07T07:40:01.1117924Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1118039Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1118460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T07:40:01.1118733Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T07:40:01.1119296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T07:40:01.1119552Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T07:40:01.1120009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T07:40:01.1120124Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T07:40:01.1120129Z 2025-09-07T07:40:01.1120224Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1120319Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1120457Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1120993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1121108Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1121534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1121616Z layer_outputs = layer( 2025-09-07T07:40:01.1122044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1122135Z attn_outputs = self.attention( 2025-09-07T07:40:01.1122558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1122671Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1123091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.1123224Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.1123640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T07:40:01.1123950Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.1124400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T07:40:01.1124607Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T07:40:01.1124613Z 2025-09-07T07:40:01.1124709Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1124809Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1124942Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1125477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1125593Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1126016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1126102Z layer_outputs = layer( 2025-09-07T07:40:01.1126524Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1126701Z attn_outputs = self.attention( 2025-09-07T07:40:01.1127137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1127252Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1127672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.1127804Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.1128225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T07:40:01.1128597Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.1129047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T07:40:01.1129143Z return torch.cat(slices, dim=3) 2025-09-07T07:40:01.1129147Z 2025-09-07T07:40:01.1129242Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1129381Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1129916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1130030Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1130451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1130528Z layer_outputs = layer( 2025-09-07T07:40:01.1130958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1131046Z attn_outputs = self.attention( 2025-09-07T07:40:01.1131473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1131588Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1132007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 541, in forward 2025-09-07T07:40:01.1132112Z value_vectors = self.value(hidden_states) 2025-09-07T07:40:01.1132116Z 2025-09-07T07:40:01.1132250Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1132786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1132902Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1133326Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1133404Z layer_outputs = layer( 2025-09-07T07:40:01.1133830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1133921Z attn_outputs = self.attention( 2025-09-07T07:40:01.1134346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1134464Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1134880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.1135011Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.1135435Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T07:40:01.1135693Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.1136233Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T07:40:01.1136443Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T07:40:01.1136449Z 2025-09-07T07:40:01.1136552Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1136684Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1137220Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1137336Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1137834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1137917Z layer_outputs = layer( 2025-09-07T07:40:01.1138344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1138437Z attn_outputs = self.attention( 2025-09-07T07:40:01.1138859Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1138974Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1139398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.1139531Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.1139952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T07:40:01.1140212Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.1140665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T07:40:01.1140757Z return torch.cat(slices, dim=3) 2025-09-07T07:40:01.1140762Z 2025-09-07T07:40:01.1140960Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1141497Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1141612Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1142036Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1142117Z layer_outputs = layer( 2025-09-07T07:40:01.1142544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1142631Z attn_outputs = self.attention( 2025-09-07T07:40:01.1143061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1143175Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1143594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T07:40:01.1143889Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T07:40:01.1144369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T07:40:01.1144493Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T07:40:01.1144503Z 2025-09-07T07:40:01.1144638Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1145270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1145409Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1145837Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1145915Z layer_outputs = layer( 2025-09-07T07:40:01.1146336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1146429Z attn_outputs = self.attention( 2025-09-07T07:40:01.1146851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1147033Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1147451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.1147592Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.1148011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T07:40:01.1148219Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T07:40:01.1148223Z 2025-09-07T07:40:01.1148364Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1148899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1149015Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1149438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1149525Z layer_outputs = layer( 2025-09-07T07:40:01.1149948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1150038Z attn_outputs = self.attention( 2025-09-07T07:40:01.1150463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1150577Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1151002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.1151134Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.1151549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T07:40:01.1151833Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T07:40:01.1152290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T07:40:01.1152496Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T07:40:01.1152502Z 2025-09-07T07:40:01.1152599Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1152735Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1153270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1153384Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1153813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1153897Z layer_outputs = layer( 2025-09-07T07:40:01.1154321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1154554Z attn_outputs = self.attention( 2025-09-07T07:40:01.1154982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1155101Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1155519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.1155655Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.1156073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T07:40:01.1156224Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T07:40:01.1156286Z 2025-09-07T07:40:01.1156423Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1156962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1157078Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1157500Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1157582Z layer_outputs = layer( 2025-09-07T07:40:01.1158003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1158095Z attn_outputs = self.attention( 2025-09-07T07:40:01.1158516Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1158635Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1159055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T07:40:01.1159188Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T07:40:01.1159608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T07:40:01.1159756Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T07:40:01.1159761Z 2025-09-07T07:40:01.1159856Z cudagraph partition due to non gpu ops 2025-09-07T07:40:01.1159995Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1160529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1160645Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1161070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1161154Z layer_outputs = layer( 2025-09-07T07:40:01.1161579Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1161667Z attn_outputs = self.attention( 2025-09-07T07:40:01.1162092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T07:40:01.1162207Z self_attention_outputs = self.self_attention( 2025-09-07T07:40:01.1162628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T07:40:01.1162929Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T07:40:01.1163424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T07:40:01.1163617Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T07:40:01.1163708Z 2025-09-07T07:40:01.1163850Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1164390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1164504Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1164932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1165010Z layer_outputs = layer( 2025-09-07T07:40:01.1165505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T07:40:01.1165729Z attn_outputs = self.attention( 2025-09-07T07:40:01.1166151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T07:40:01.1166352Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T07:40:01.1166774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T07:40:01.1166882Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.1166887Z 2025-09-07T07:40:01.1167019Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1167555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1167671Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1168100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1168183Z layer_outputs = layer( 2025-09-07T07:40:01.1168610Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.1168777Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.1169198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T07:40:01.1169290Z return apply_chunking_to_forward( 2025-09-07T07:40:01.1169693Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:01.1169785Z return forward_fn(*input_tensors) 2025-09-07T07:40:01.1170236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T07:40:01.1170342Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.1170761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T07:40:01.1170868Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.1170872Z 2025-09-07T07:40:01.1171006Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1171543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1171655Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1172080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1172158Z layer_outputs = layer( 2025-09-07T07:40:01.1172580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.1172748Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.1173284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T07:40:01.1173386Z return apply_chunking_to_forward( 2025-09-07T07:40:01.1173790Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:01.1173881Z return forward_fn(*input_tensors) 2025-09-07T07:40:01.1174329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T07:40:01.1174429Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.1174855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T07:40:01.1175025Z hidden_states = self.act_fn(hidden_states) 2025-09-07T07:40:01.1175030Z 2025-09-07T07:40:01.1175166Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1175706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1175819Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1176245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1176322Z layer_outputs = layer( 2025-09-07T07:40:01.1176751Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.1176916Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.1177348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T07:40:01.1177439Z return apply_chunking_to_forward( 2025-09-07T07:40:01.1177844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:01.1177938Z return forward_fn(*input_tensors) 2025-09-07T07:40:01.1178386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T07:40:01.1178481Z return self.output(hidden_states) 2025-09-07T07:40:01.1178902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T07:40:01.1179001Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:01.1179009Z 2025-09-07T07:40:01.1179142Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1179681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1179797Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1180222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T07:40:01.1180302Z layer_outputs = layer( 2025-09-07T07:40:01.1180725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T07:40:01.1180886Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T07:40:01.1180891Z 2025-09-07T07:40:01.1181029Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:01.1181563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T07:40:01.1181682Z hidden_states = _ReversibleFunction.apply( 2025-09-07T07:40:01.1182105Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1753, in forward 2025-09-07T07:40:01.1182315Z return torch.cat([attn_output, hidden_states], dim=-1) 2025-09-07T07:40:01.1182321Z 2025-09-07T07:40:04.9518793Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:04.9519618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 2495, in torch_dynamo_resume_in_forward_at_2481 2025-09-07T07:40:04.9520367Z logits = self.lm_head(sequence_output) 2025-09-07T07:40:04.9520994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1886, in forward 2025-09-07T07:40:04.9521860Z return apply_chunking_to_forward(self.forward_chunk, self.chunk_size_lm_head, self.seq_len_dim, hidden_states) 2025-09-07T07:40:04.9523135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:04.9523742Z return forward_fn(*input_tensors) 2025-09-07T07:40:04.9524394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1889, in forward_chunk 2025-09-07T07:40:04.9525062Z hidden_states = self.decoder(hidden_states) 2025-09-07T07:40:04.9525272Z 2025-09-07T07:40:05.1502111Z pass 2025-09-07T07:40:05.1502527Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:40:06.7429169Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:40:06.7430367Z import pynvml # type: ignore[import] 2025-09-07T07:40:09.3263205Z 2025-09-07T07:40:19.3489596Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:40:19.3489910Z loading model: 0it [00:10, ?it/s] 2025-09-07T07:40:19.3710396Z cpu eval hf_Roberta_base 2025-09-07T07:40:25.1819757Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:40:26.1437799Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:40:27.0904094Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:40:37.8865115Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.8865738Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.8866216Z return mod(**inputs) 2025-09-07T07:40:37.8866818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.8867481Z outputs = self.roberta( 2025-09-07T07:40:37.8868085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 817, in forward 2025-09-07T07:40:37.8868728Z embedding_output = self.embeddings( 2025-09-07T07:40:37.8869354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 90, in forward 2025-09-07T07:40:37.8870189Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length) 2025-09-07T07:40:37.8871122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1576, in create_position_ids_from_input_ids 2025-09-07T07:40:37.8871922Z mask = input_ids.ne(padding_idx).int() 2025-09-07T07:40:37.8872120Z 2025-09-07T07:40:37.8872225Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.8872515Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.8872793Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.8873072Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.8873732Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.8874035Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.8874312Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.8874588Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.8874865Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.8875140Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.8875419Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.8875697Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.8876012Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.8876536Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.8877007Z return mod(**inputs) 2025-09-07T07:40:37.8877803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.8878443Z outputs = self.roberta( 2025-09-07T07:40:37.8879040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 817, in forward 2025-09-07T07:40:37.8879669Z embedding_output = self.embeddings( 2025-09-07T07:40:37.8880292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 90, in forward 2025-09-07T07:40:37.8881119Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length) 2025-09-07T07:40:37.8882044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1577, in create_position_ids_from_input_ids 2025-09-07T07:40:37.8882953Z incremental_indices = (torch.cumsum(mask, dim=1).type_as(mask) + past_key_values_length) * mask 2025-09-07T07:40:37.8883330Z 2025-09-07T07:40:37.8883468Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.8883997Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.8884464Z return mod(**inputs) 2025-09-07T07:40:37.8885047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.8885673Z outputs = self.roberta( 2025-09-07T07:40:37.8886257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 817, in forward 2025-09-07T07:40:37.8886893Z embedding_output = self.embeddings( 2025-09-07T07:40:37.8887510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 90, in forward 2025-09-07T07:40:37.8888339Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length) 2025-09-07T07:40:37.8889272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1577, in create_position_ids_from_input_ids 2025-09-07T07:40:37.8890175Z incremental_indices = (torch.cumsum(mask, dim=1).type_as(mask) + past_key_values_length) * mask 2025-09-07T07:40:37.8890542Z 2025-09-07T07:40:37.8890677Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.8891206Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.8891675Z return mod(**inputs) 2025-09-07T07:40:37.8892257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.8892881Z outputs = self.roberta( 2025-09-07T07:40:37.8893462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 817, in forward 2025-09-07T07:40:37.8894102Z embedding_output = self.embeddings( 2025-09-07T07:40:37.8894841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 120, in forward 2025-09-07T07:40:37.8895489Z embeddings = self.LayerNorm(embeddings) 2025-09-07T07:40:37.8895687Z 2025-09-07T07:40:37.8895824Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.8896345Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.8896816Z return mod(**inputs) 2025-09-07T07:40:37.8897395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.8898016Z outputs = self.roberta( 2025-09-07T07:40:37.8898603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.8899319Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.8899928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.8900548Z layer_outputs = layer_module( 2025-09-07T07:40:37.8901054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.8901594Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.8902219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.8902854Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.8903436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.8903999Z return func(*args, **kwargs) 2025-09-07T07:40:37.8904598Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.8905215Z self_outputs = self.self( 2025-09-07T07:40:37.8905755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.8906317Z return func(*args, **kwargs) 2025-09-07T07:40:37.8906914Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T07:40:37.8907756Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T07:40:37.8908164Z 2025-09-07T07:40:37.8908300Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.8908826Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.8909294Z return mod(**inputs) 2025-09-07T07:40:37.8909876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.8910498Z outputs = self.roberta( 2025-09-07T07:40:37.8911084Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.8911704Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.8912306Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.8912923Z layer_outputs = layer_module( 2025-09-07T07:40:37.8913431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.8913969Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.8914592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.8915231Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.8915929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.8916503Z return func(*args, **kwargs) 2025-09-07T07:40:37.8917095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.8917706Z self_outputs = self.self( 2025-09-07T07:40:37.8918238Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.8918794Z return func(*args, **kwargs) 2025-09-07T07:40:37.8919387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T07:40:37.8920130Z self.key(current_states) 2025-09-07T07:40:37.8920283Z 2025-09-07T07:40:37.8920418Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.8920937Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.8921402Z return mod(**inputs) 2025-09-07T07:40:37.8921982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.8922600Z outputs = self.roberta( 2025-09-07T07:40:37.8923176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.8923796Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.8924403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.8925018Z layer_outputs = layer_module( 2025-09-07T07:40:37.8925534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.8926069Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.8926695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.8927334Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.8927924Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.8928485Z return func(*args, **kwargs) 2025-09-07T07:40:37.8929079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.8929693Z self_outputs = self.self( 2025-09-07T07:40:37.8930219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.8930785Z return func(*args, **kwargs) 2025-09-07T07:40:37.8931374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T07:40:37.8931990Z self.value(current_states) 2025-09-07T07:40:37.8932149Z 2025-09-07T07:40:37.8932249Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.8932565Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.8933082Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.8933546Z return mod(**inputs) 2025-09-07T07:40:37.8934124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.8934739Z outputs = self.roberta( 2025-09-07T07:40:37.8935319Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.8935945Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.8936552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.8937289Z layer_outputs = layer_module( 2025-09-07T07:40:37.8937808Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.8938348Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.8938971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.8939608Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.8940182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.8940738Z return func(*args, **kwargs) 2025-09-07T07:40:37.8941423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.8942035Z self_outputs = self.self( 2025-09-07T07:40:37.8942571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.8943126Z return func(*args, **kwargs) 2025-09-07T07:40:37.8943719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T07:40:37.8944422Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T07:40:37.8944695Z 2025-09-07T07:40:37.8944834Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.8945350Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.8945807Z return mod(**inputs) 2025-09-07T07:40:37.8946390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.8947009Z outputs = self.roberta( 2025-09-07T07:40:37.8947592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.8948209Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.8948809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.8949429Z layer_outputs = layer_module( 2025-09-07T07:40:37.8949937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.8950478Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.8951095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.8951736Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.8952311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.8952873Z return func(*args, **kwargs) 2025-09-07T07:40:37.8953470Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T07:40:37.8954168Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T07:40:37.8954872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T07:40:37.8955509Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.8955707Z 2025-09-07T07:40:37.8955845Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.8956364Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.8956828Z return mod(**inputs) 2025-09-07T07:40:37.8957409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.8958147Z outputs = self.roberta( 2025-09-07T07:40:37.8958743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.8959360Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.8959960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.8960572Z layer_outputs = layer_module( 2025-09-07T07:40:37.8961078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.8961616Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.8962315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.8962948Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.8963554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.8964149Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.8964805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.8965624Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.8966316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T07:40:37.8966952Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.8967148Z 2025-09-07T07:40:37.8967290Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.8967808Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.8968266Z return mod(**inputs) 2025-09-07T07:40:37.8968849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.8969471Z outputs = self.roberta( 2025-09-07T07:40:37.8970053Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.8970670Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.8971267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.8971881Z layer_outputs = layer_module( 2025-09-07T07:40:37.8972388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.8972925Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.8973546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.8974180Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.8974784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.8975377Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.8976030Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.8976764Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.8977970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T07:40:37.8978652Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T07:40:37.8979220Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T07:40:37.8979930Z return self.act(input) 2025-09-07T07:40:37.8980113Z 2025-09-07T07:40:37.8980262Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.8980781Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.8981243Z return mod(**inputs) 2025-09-07T07:40:37.8981820Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.8982441Z outputs = self.roberta( 2025-09-07T07:40:37.8983017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.8983769Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.8984374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.8984987Z layer_outputs = layer_module( 2025-09-07T07:40:37.8985493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.8986029Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.8986651Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.8987283Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.8987882Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.8988470Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.8989121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T07:40:37.8989881Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T07:40:37.8990599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T07:40:37.8991231Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.8991430Z 2025-09-07T07:40:37.8991564Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.8992081Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.8992542Z return mod(**inputs) 2025-09-07T07:40:37.8993118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.8993738Z outputs = self.roberta( 2025-09-07T07:40:37.8994316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.8994940Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.8995551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.8996169Z layer_outputs = layer_module( 2025-09-07T07:40:37.8996672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.8997210Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.8997833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.8998467Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.8999040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.8999602Z return func(*args, **kwargs) 2025-09-07T07:40:37.9000197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9000917Z self_outputs = self.self( 2025-09-07T07:40:37.9001469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9002021Z return func(*args, **kwargs) 2025-09-07T07:40:37.9002608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T07:40:37.9003442Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T07:40:37.9003849Z 2025-09-07T07:40:37.9003984Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9004499Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9005047Z return mod(**inputs) 2025-09-07T07:40:37.9005620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9006245Z outputs = self.roberta( 2025-09-07T07:40:37.9006831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9007450Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9008048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9008666Z layer_outputs = layer_module( 2025-09-07T07:40:37.9009172Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9009710Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9010337Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9010968Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9011543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9012099Z return func(*args, **kwargs) 2025-09-07T07:40:37.9012693Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9013305Z self_outputs = self.self( 2025-09-07T07:40:37.9013830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9014387Z return func(*args, **kwargs) 2025-09-07T07:40:37.9014977Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T07:40:37.9015591Z self.key(current_states) 2025-09-07T07:40:37.9015742Z 2025-09-07T07:40:37.9015877Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9016397Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9016858Z return mod(**inputs) 2025-09-07T07:40:37.9017434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9018053Z outputs = self.roberta( 2025-09-07T07:40:37.9018630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9019254Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9019859Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9020480Z layer_outputs = layer_module( 2025-09-07T07:40:37.9020988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9021524Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9022264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9022916Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9023487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9024038Z return func(*args, **kwargs) 2025-09-07T07:40:37.9024629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9025238Z self_outputs = self.self( 2025-09-07T07:40:37.9025769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9026400Z return func(*args, **kwargs) 2025-09-07T07:40:37.9026989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T07:40:37.9027602Z self.value(current_states) 2025-09-07T07:40:37.9027760Z 2025-09-07T07:40:37.9027855Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.9028172Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9028688Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9029143Z return mod(**inputs) 2025-09-07T07:40:37.9029718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9030336Z outputs = self.roberta( 2025-09-07T07:40:37.9030911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9031526Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9032133Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9032750Z layer_outputs = layer_module( 2025-09-07T07:40:37.9033271Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9033808Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9034425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9035063Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9035636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9036199Z return func(*args, **kwargs) 2025-09-07T07:40:37.9036785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9037398Z self_outputs = self.self( 2025-09-07T07:40:37.9037929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9038484Z return func(*args, **kwargs) 2025-09-07T07:40:37.9039076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T07:40:37.9039781Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T07:40:37.9040056Z 2025-09-07T07:40:37.9040190Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9040703Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9041169Z return mod(**inputs) 2025-09-07T07:40:37.9041745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9042363Z outputs = self.roberta( 2025-09-07T07:40:37.9043061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9043689Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9044295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9044909Z layer_outputs = layer_module( 2025-09-07T07:40:37.9045413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9045945Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9046568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9047287Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9047862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9059746Z return func(*args, **kwargs) 2025-09-07T07:40:37.9060479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T07:40:37.9061189Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T07:40:37.9061896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T07:40:37.9062534Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9062737Z 2025-09-07T07:40:37.9062875Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9063413Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9063879Z return mod(**inputs) 2025-09-07T07:40:37.9064464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9065086Z outputs = self.roberta( 2025-09-07T07:40:37.9065750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9066379Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9066988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9067608Z layer_outputs = layer_module( 2025-09-07T07:40:37.9068113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9068651Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9069277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9069911Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9070515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9071107Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9071766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9072506Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9073200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T07:40:37.9073832Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9074038Z 2025-09-07T07:40:37.9074174Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9074695Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9075158Z return mod(**inputs) 2025-09-07T07:40:37.9075969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9076610Z outputs = self.roberta( 2025-09-07T07:40:37.9077193Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9077812Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9078417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9079031Z layer_outputs = layer_module( 2025-09-07T07:40:37.9079533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9080214Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9080841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9081473Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9082074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9082664Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9083322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9084057Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9084745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T07:40:37.9085419Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T07:40:37.9085990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T07:40:37.9086498Z return self.act(input) 2025-09-07T07:40:37.9086643Z 2025-09-07T07:40:37.9086784Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9087302Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9087758Z return mod(**inputs) 2025-09-07T07:40:37.9088334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9088952Z outputs = self.roberta( 2025-09-07T07:40:37.9089535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9090158Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9090757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9091379Z layer_outputs = layer_module( 2025-09-07T07:40:37.9091885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9092423Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9093041Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9093675Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9094276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9094867Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9095525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T07:40:37.9096274Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T07:40:37.9097123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T07:40:37.9097766Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9097961Z 2025-09-07T07:40:37.9098098Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9098620Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9099076Z return mod(**inputs) 2025-09-07T07:40:37.9099657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9100273Z outputs = self.roberta( 2025-09-07T07:40:37.9100932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9101551Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9102156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9102775Z layer_outputs = layer_module( 2025-09-07T07:40:37.9103284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9103821Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9104439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9105076Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9105651Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9106218Z return func(*args, **kwargs) 2025-09-07T07:40:37.9106812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9107423Z self_outputs = self.self( 2025-09-07T07:40:37.9107957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9108511Z return func(*args, **kwargs) 2025-09-07T07:40:37.9109102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T07:40:37.9109938Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T07:40:37.9110343Z 2025-09-07T07:40:37.9110478Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9110997Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9111466Z return mod(**inputs) 2025-09-07T07:40:37.9112053Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9112667Z outputs = self.roberta( 2025-09-07T07:40:37.9113245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9113863Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9114468Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9115083Z layer_outputs = layer_module( 2025-09-07T07:40:37.9115584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9116120Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9116744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9117379Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9118070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9118633Z return func(*args, **kwargs) 2025-09-07T07:40:37.9119226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9119836Z self_outputs = self.self( 2025-09-07T07:40:37.9120365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9120918Z return func(*args, **kwargs) 2025-09-07T07:40:37.9121504Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T07:40:37.9122193Z self.key(current_states) 2025-09-07T07:40:37.9122343Z 2025-09-07T07:40:37.9122482Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9123003Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9123460Z return mod(**inputs) 2025-09-07T07:40:37.9124037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9124655Z outputs = self.roberta( 2025-09-07T07:40:37.9125232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9125848Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9126448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9127067Z layer_outputs = layer_module( 2025-09-07T07:40:37.9127573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9128116Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9128734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9129367Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9129942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9130499Z return func(*args, **kwargs) 2025-09-07T07:40:37.9131091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9131698Z self_outputs = self.self( 2025-09-07T07:40:37.9132234Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9132789Z return func(*args, **kwargs) 2025-09-07T07:40:37.9133388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T07:40:37.9134003Z self.value(current_states) 2025-09-07T07:40:37.9134162Z 2025-09-07T07:40:37.9134259Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.9134580Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9135097Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9135558Z return mod(**inputs) 2025-09-07T07:40:37.9136135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9136756Z outputs = self.roberta( 2025-09-07T07:40:37.9137343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9137965Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9138708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9139332Z layer_outputs = layer_module( 2025-09-07T07:40:37.9139844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9140379Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9141000Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9141634Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9142202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9142840Z return func(*args, **kwargs) 2025-09-07T07:40:37.9143433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9144052Z self_outputs = self.self( 2025-09-07T07:40:37.9144580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9145135Z return func(*args, **kwargs) 2025-09-07T07:40:37.9145724Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T07:40:37.9146427Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T07:40:37.9146697Z 2025-09-07T07:40:37.9146833Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9147343Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9147807Z return mod(**inputs) 2025-09-07T07:40:37.9148382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9149001Z outputs = self.roberta( 2025-09-07T07:40:37.9149585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9150199Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9150804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9151418Z layer_outputs = layer_module( 2025-09-07T07:40:37.9151923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9152454Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9153075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9153717Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9154295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9154855Z return func(*args, **kwargs) 2025-09-07T07:40:37.9155441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T07:40:37.9156136Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T07:40:37.9156831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T07:40:37.9157464Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9157662Z 2025-09-07T07:40:37.9157799Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9158315Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9158775Z return mod(**inputs) 2025-09-07T07:40:37.9159471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9160107Z outputs = self.roberta( 2025-09-07T07:40:37.9160686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9161298Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9161902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9162516Z layer_outputs = layer_module( 2025-09-07T07:40:37.9163022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9163640Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9164264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9164897Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9165576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9166173Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9166823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9167557Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9168247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T07:40:37.9168885Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9169082Z 2025-09-07T07:40:37.9169220Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9169733Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9170195Z return mod(**inputs) 2025-09-07T07:40:37.9170770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9171388Z outputs = self.roberta( 2025-09-07T07:40:37.9171966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9172578Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9173182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9173796Z layer_outputs = layer_module( 2025-09-07T07:40:37.9174310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9174842Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9175466Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9176101Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9176704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9177299Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9177951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9178683Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9179375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T07:40:37.9180053Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T07:40:37.9180824Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T07:40:37.9181338Z return self.act(input) 2025-09-07T07:40:37.9181489Z 2025-09-07T07:40:37.9181623Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9182141Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9182598Z return mod(**inputs) 2025-09-07T07:40:37.9183176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9183789Z outputs = self.roberta( 2025-09-07T07:40:37.9184370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9185121Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9185725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9186338Z layer_outputs = layer_module( 2025-09-07T07:40:37.9186845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9187379Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9187999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9188626Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9189223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9189812Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9190470Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T07:40:37.9191225Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T07:40:37.9191932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T07:40:37.9192559Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9192758Z 2025-09-07T07:40:37.9192893Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9193407Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9193866Z return mod(**inputs) 2025-09-07T07:40:37.9194443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9195061Z outputs = self.roberta( 2025-09-07T07:40:37.9195640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9196257Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9196866Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9197475Z layer_outputs = layer_module( 2025-09-07T07:40:37.9197982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9198518Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9199138Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9199769Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9200336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9200896Z return func(*args, **kwargs) 2025-09-07T07:40:37.9201601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9202226Z self_outputs = self.self( 2025-09-07T07:40:37.9202758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9203312Z return func(*args, **kwargs) 2025-09-07T07:40:37.9203903Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T07:40:37.9204734Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T07:40:37.9205134Z 2025-09-07T07:40:37.9205275Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9205876Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9206331Z return mod(**inputs) 2025-09-07T07:40:37.9206910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9207528Z outputs = self.roberta( 2025-09-07T07:40:37.9208104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9208716Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9209320Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9209931Z layer_outputs = layer_module( 2025-09-07T07:40:37.9210433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9210970Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9211585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9212223Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9212796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9213348Z return func(*args, **kwargs) 2025-09-07T07:40:37.9213935Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9214541Z self_outputs = self.self( 2025-09-07T07:40:37.9215074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9215631Z return func(*args, **kwargs) 2025-09-07T07:40:37.9216219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T07:40:37.9216829Z self.key(current_states) 2025-09-07T07:40:37.9216981Z 2025-09-07T07:40:37.9217115Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9217635Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9218095Z return mod(**inputs) 2025-09-07T07:40:37.9218669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9219281Z outputs = self.roberta( 2025-09-07T07:40:37.9219861Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9220478Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9221086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9221707Z layer_outputs = layer_module( 2025-09-07T07:40:37.9222211Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9222864Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9223501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9224136Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9224702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9225255Z return func(*args, **kwargs) 2025-09-07T07:40:37.9225846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9226461Z self_outputs = self.self( 2025-09-07T07:40:37.9227065Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9227617Z return func(*args, **kwargs) 2025-09-07T07:40:37.9228211Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T07:40:37.9228826Z self.value(current_states) 2025-09-07T07:40:37.9228983Z 2025-09-07T07:40:37.9229083Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.9229403Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9229917Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9230378Z return mod(**inputs) 2025-09-07T07:40:37.9230957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9231575Z outputs = self.roberta( 2025-09-07T07:40:37.9232156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9232772Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9233379Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9233996Z layer_outputs = layer_module( 2025-09-07T07:40:37.9234503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9235033Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9235654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9236290Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9236862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9237420Z return func(*args, **kwargs) 2025-09-07T07:40:37.9238010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9238624Z self_outputs = self.self( 2025-09-07T07:40:37.9239152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9239705Z return func(*args, **kwargs) 2025-09-07T07:40:37.9240298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T07:40:37.9241002Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T07:40:37.9241277Z 2025-09-07T07:40:37.9241414Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9241929Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9242396Z return mod(**inputs) 2025-09-07T07:40:37.9242969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9243699Z outputs = self.roberta( 2025-09-07T07:40:37.9244293Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9244912Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9245515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9246124Z layer_outputs = layer_module( 2025-09-07T07:40:37.9246636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9247171Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9247896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9248528Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9249102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9249660Z return func(*args, **kwargs) 2025-09-07T07:40:37.9250254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T07:40:37.9250947Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T07:40:37.9251642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T07:40:37.9252274Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9252475Z 2025-09-07T07:40:37.9252614Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9253129Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9253586Z return mod(**inputs) 2025-09-07T07:40:37.9254159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9254776Z outputs = self.roberta( 2025-09-07T07:40:37.9255353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9255973Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9256578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9257203Z layer_outputs = layer_module( 2025-09-07T07:40:37.9257713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9258257Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9258878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9259517Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9260119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9260711Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9261362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9262103Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9262791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T07:40:37.9263427Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9263624Z 2025-09-07T07:40:37.9263762Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9264393Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9264861Z return mod(**inputs) 2025-09-07T07:40:37.9265506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9266130Z outputs = self.roberta( 2025-09-07T07:40:37.9266710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9267324Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9267930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9268695Z layer_outputs = layer_module( 2025-09-07T07:40:37.9269200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9269733Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9270355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9270986Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9271586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9272176Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9272828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9273563Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9274253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T07:40:37.9274927Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T07:40:37.9275495Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T07:40:37.9275999Z return self.act(input) 2025-09-07T07:40:37.9276145Z 2025-09-07T07:40:37.9276279Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9276793Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9277253Z return mod(**inputs) 2025-09-07T07:40:37.9277829Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9278445Z outputs = self.roberta( 2025-09-07T07:40:37.9279026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9279651Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9280257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9280870Z layer_outputs = layer_module( 2025-09-07T07:40:37.9281376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9281913Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9282535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9283168Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9283772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9284368Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9285024Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T07:40:37.9285927Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T07:40:37.9286657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T07:40:37.9287286Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9287489Z 2025-09-07T07:40:37.9287627Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9288141Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9288602Z return mod(**inputs) 2025-09-07T07:40:37.9289181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9289871Z outputs = self.roberta( 2025-09-07T07:40:37.9290449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9291066Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9291668Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9292277Z layer_outputs = layer_module( 2025-09-07T07:40:37.9292783Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9293321Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9293942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9294572Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9295143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9295701Z return func(*args, **kwargs) 2025-09-07T07:40:37.9296293Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9296907Z self_outputs = self.self( 2025-09-07T07:40:37.9297436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9297985Z return func(*args, **kwargs) 2025-09-07T07:40:37.9298576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T07:40:37.9299408Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T07:40:37.9299811Z 2025-09-07T07:40:37.9299950Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9300469Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9300924Z return mod(**inputs) 2025-09-07T07:40:37.9301505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9302123Z outputs = self.roberta( 2025-09-07T07:40:37.9302699Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9303312Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9303914Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9304530Z layer_outputs = layer_module( 2025-09-07T07:40:37.9305034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9305573Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9306189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9306922Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9307506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9308062Z return func(*args, **kwargs) 2025-09-07T07:40:37.9308651Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9309260Z self_outputs = self.self( 2025-09-07T07:40:37.9309794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9310347Z return func(*args, **kwargs) 2025-09-07T07:40:37.9311042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T07:40:37.9311651Z self.key(current_states) 2025-09-07T07:40:37.9311804Z 2025-09-07T07:40:37.9311944Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9312467Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9312927Z return mod(**inputs) 2025-09-07T07:40:37.9313503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9314119Z outputs = self.roberta( 2025-09-07T07:40:37.9314700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9315316Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9315916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9316537Z layer_outputs = layer_module( 2025-09-07T07:40:37.9317041Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9317577Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9318200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9318835Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9319402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9319957Z return func(*args, **kwargs) 2025-09-07T07:40:37.9320552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9321170Z self_outputs = self.self( 2025-09-07T07:40:37.9321698Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9322250Z return func(*args, **kwargs) 2025-09-07T07:40:37.9322847Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T07:40:37.9323461Z self.value(current_states) 2025-09-07T07:40:37.9323616Z 2025-09-07T07:40:37.9323715Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.9324034Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9324544Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9325005Z return mod(**inputs) 2025-09-07T07:40:37.9325584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9326206Z outputs = self.roberta( 2025-09-07T07:40:37.9326783Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9327400Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9328122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9328746Z layer_outputs = layer_module( 2025-09-07T07:40:37.9329252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9329783Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9330405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9331036Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9331606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9332265Z return func(*args, **kwargs) 2025-09-07T07:40:37.9332860Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9333473Z self_outputs = self.self( 2025-09-07T07:40:37.9334004Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9334557Z return func(*args, **kwargs) 2025-09-07T07:40:37.9335142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T07:40:37.9335844Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T07:40:37.9336119Z 2025-09-07T07:40:37.9336255Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9336774Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9337236Z return mod(**inputs) 2025-09-07T07:40:37.9337812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9338430Z outputs = self.roberta( 2025-09-07T07:40:37.9339009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9339627Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9340229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9340839Z layer_outputs = layer_module( 2025-09-07T07:40:37.9341344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9341877Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9342498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9343129Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9343702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9344259Z return func(*args, **kwargs) 2025-09-07T07:40:37.9344850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T07:40:37.9345544Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T07:40:37.9346235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T07:40:37.9346869Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9347077Z 2025-09-07T07:40:37.9347214Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9347731Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9348189Z return mod(**inputs) 2025-09-07T07:40:37.9348863Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9349491Z outputs = self.roberta( 2025-09-07T07:40:37.9350070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9350690Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9351291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9351900Z layer_outputs = layer_module( 2025-09-07T07:40:37.9352405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9353025Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9353649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9354282Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9354878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9355467Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9356122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9356855Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9357542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T07:40:37.9358182Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9358383Z 2025-09-07T07:40:37.9358518Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9359037Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9359495Z return mod(**inputs) 2025-09-07T07:40:37.9360069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9360688Z outputs = self.roberta( 2025-09-07T07:40:37.9361265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9361881Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9362485Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9363108Z layer_outputs = layer_module( 2025-09-07T07:40:37.9363626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9364169Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9364787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9365510Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9366124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9366717Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9367372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9368106Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9368800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T07:40:37.9369474Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T07:40:37.9370226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T07:40:37.9370745Z return self.act(input) 2025-09-07T07:40:37.9370891Z 2025-09-07T07:40:37.9371028Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9371543Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9372003Z return mod(**inputs) 2025-09-07T07:40:37.9372580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9373198Z outputs = self.roberta( 2025-09-07T07:40:37.9373910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9374534Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9375144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9375759Z layer_outputs = layer_module( 2025-09-07T07:40:37.9376268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9376800Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9377418Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9378054Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9378656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9379256Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9379906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T07:40:37.9380669Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T07:40:37.9381383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T07:40:37.9382015Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9382212Z 2025-09-07T07:40:37.9382351Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9382863Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9383322Z return mod(**inputs) 2025-09-07T07:40:37.9383897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9384520Z outputs = self.roberta( 2025-09-07T07:40:37.9385096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9385717Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9386322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9386934Z layer_outputs = layer_module( 2025-09-07T07:40:37.9387442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9387976Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9388596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9389233Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9389813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9390371Z return func(*args, **kwargs) 2025-09-07T07:40:37.9391089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9391715Z self_outputs = self.self( 2025-09-07T07:40:37.9392244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9392799Z return func(*args, **kwargs) 2025-09-07T07:40:37.9393388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T07:40:37.9394222Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T07:40:37.9394702Z 2025-09-07T07:40:37.9394839Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9395354Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9395814Z return mod(**inputs) 2025-09-07T07:40:37.9396390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9397008Z outputs = self.roberta( 2025-09-07T07:40:37.9397588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9398204Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9398807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9399418Z layer_outputs = layer_module( 2025-09-07T07:40:37.9399924Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9400467Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9401092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9401725Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9402295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9402853Z return func(*args, **kwargs) 2025-09-07T07:40:37.9403443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9404052Z self_outputs = self.self( 2025-09-07T07:40:37.9404580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9405134Z return func(*args, **kwargs) 2025-09-07T07:40:37.9405734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T07:40:37.9406345Z self.key(current_states) 2025-09-07T07:40:37.9406495Z 2025-09-07T07:40:37.9406637Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9407148Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9407605Z return mod(**inputs) 2025-09-07T07:40:37.9408179Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9408796Z outputs = self.roberta( 2025-09-07T07:40:37.9409371Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9409987Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9410596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9411216Z layer_outputs = layer_module( 2025-09-07T07:40:37.9411854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9412407Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9413024Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9413657Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9414228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9414784Z return func(*args, **kwargs) 2025-09-07T07:40:37.9415372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9416053Z self_outputs = self.self( 2025-09-07T07:40:37.9416581Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9417142Z return func(*args, **kwargs) 2025-09-07T07:40:37.9417734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T07:40:37.9418344Z self.value(current_states) 2025-09-07T07:40:37.9418509Z 2025-09-07T07:40:37.9418607Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.9418925Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9419439Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9419891Z return mod(**inputs) 2025-09-07T07:40:37.9420465Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9421083Z outputs = self.roberta( 2025-09-07T07:40:37.9421662Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9422282Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9422881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9423498Z layer_outputs = layer_module( 2025-09-07T07:40:37.9424008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9424540Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9425158Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9425787Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9426367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9426923Z return func(*args, **kwargs) 2025-09-07T07:40:37.9427520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9428132Z self_outputs = self.self( 2025-09-07T07:40:37.9428661Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9429214Z return func(*args, **kwargs) 2025-09-07T07:40:37.9429805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T07:40:37.9430513Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T07:40:37.9430784Z 2025-09-07T07:40:37.9430918Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9431443Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9431904Z return mod(**inputs) 2025-09-07T07:40:37.9432601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9433230Z outputs = self.roberta( 2025-09-07T07:40:37.9433808Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9434425Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9435026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9435640Z layer_outputs = layer_module( 2025-09-07T07:40:37.9436141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9436762Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9437381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9438016Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9438586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9439139Z return func(*args, **kwargs) 2025-09-07T07:40:37.9439728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T07:40:37.9440422Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T07:40:37.9441118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T07:40:37.9441753Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9441954Z 2025-09-07T07:40:37.9442089Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9442604Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9443064Z return mod(**inputs) 2025-09-07T07:40:37.9443639Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9444256Z outputs = self.roberta( 2025-09-07T07:40:37.9444831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9445450Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9446056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9446672Z layer_outputs = layer_module( 2025-09-07T07:40:37.9447180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9447715Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9448338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9448974Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9449575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9450161Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9450818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9451553Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9452243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T07:40:37.9452880Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9453078Z 2025-09-07T07:40:37.9453212Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9453840Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9454316Z return mod(**inputs) 2025-09-07T07:40:37.9454890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9455505Z outputs = self.roberta( 2025-09-07T07:40:37.9456083Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9456700Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9457303Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9458000Z layer_outputs = layer_module( 2025-09-07T07:40:37.9458500Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9459043Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9459661Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9460295Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9460897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9461482Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9462135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9462878Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9463565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T07:40:37.9464241Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T07:40:37.9464799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T07:40:37.9465383Z return self.act(input) 2025-09-07T07:40:37.9465538Z 2025-09-07T07:40:37.9465680Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9466199Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9466657Z return mod(**inputs) 2025-09-07T07:40:37.9467230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9467850Z outputs = self.roberta( 2025-09-07T07:40:37.9468434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9469051Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9469657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9470286Z layer_outputs = layer_module( 2025-09-07T07:40:37.9470792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9471328Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9471947Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9472052Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9472457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9472560Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9473240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T07:40:37.9473431Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T07:40:37.9473863Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T07:40:37.9473967Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9473972Z 2025-09-07T07:40:37.9474108Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9474396Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9474479Z return mod(**inputs) 2025-09-07T07:40:37.9474918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9475119Z outputs = self.roberta( 2025-09-07T07:40:37.9475556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9475647Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9476080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9476167Z layer_outputs = layer_module( 2025-09-07T07:40:37.9476514Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9476609Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9477043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9477155Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9477531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9477620Z return func(*args, **kwargs) 2025-09-07T07:40:37.9478054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9478141Z self_outputs = self.self( 2025-09-07T07:40:37.9478518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9478602Z return func(*args, **kwargs) 2025-09-07T07:40:37.9479040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T07:40:37.9479347Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T07:40:37.9479356Z 2025-09-07T07:40:37.9479496Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9479784Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9479866Z return mod(**inputs) 2025-09-07T07:40:37.9480302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9480384Z outputs = self.roberta( 2025-09-07T07:40:37.9480817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9480906Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9481342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9481427Z layer_outputs = layer_module( 2025-09-07T07:40:37.9481772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9481872Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9482422Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9482539Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9482915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9482998Z return func(*args, **kwargs) 2025-09-07T07:40:37.9483431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9483515Z self_outputs = self.self( 2025-09-07T07:40:37.9494631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9494991Z return func(*args, **kwargs) 2025-09-07T07:40:37.9495475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T07:40:37.9495566Z self.key(current_states) 2025-09-07T07:40:37.9495577Z 2025-09-07T07:40:37.9495722Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9496028Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9496106Z return mod(**inputs) 2025-09-07T07:40:37.9496567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9496651Z outputs = self.roberta( 2025-09-07T07:40:37.9497102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9497197Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9497647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9497738Z layer_outputs = layer_module( 2025-09-07T07:40:37.9498096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9498202Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9498648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9498760Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9499146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9499234Z return func(*args, **kwargs) 2025-09-07T07:40:37.9499677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9499765Z self_outputs = self.self( 2025-09-07T07:40:37.9500154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9500241Z return func(*args, **kwargs) 2025-09-07T07:40:37.9500686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T07:40:37.9500774Z self.value(current_states) 2025-09-07T07:40:37.9500779Z 2025-09-07T07:40:37.9500881Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.9501024Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9501321Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9501401Z return mod(**inputs) 2025-09-07T07:40:37.9501853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9501943Z outputs = self.roberta( 2025-09-07T07:40:37.9502492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9502590Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9503044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9503132Z layer_outputs = layer_module( 2025-09-07T07:40:37.9503488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9503587Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9504029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9504215Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9504593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9504679Z return func(*args, **kwargs) 2025-09-07T07:40:37.9505113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9505197Z self_outputs = self.self( 2025-09-07T07:40:37.9505575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9505657Z return func(*args, **kwargs) 2025-09-07T07:40:37.9506089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T07:40:37.9506265Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T07:40:37.9506270Z 2025-09-07T07:40:37.9506412Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9506700Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9506777Z return mod(**inputs) 2025-09-07T07:40:37.9507222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9507305Z outputs = self.roberta( 2025-09-07T07:40:37.9507741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9507831Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9508261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9508350Z layer_outputs = layer_module( 2025-09-07T07:40:37.9508690Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9508793Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9509226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9509336Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9509711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9509792Z return func(*args, **kwargs) 2025-09-07T07:40:37.9510228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T07:40:37.9510395Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T07:40:37.9510828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T07:40:37.9510936Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9510941Z 2025-09-07T07:40:37.9511074Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9511467Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9511548Z return mod(**inputs) 2025-09-07T07:40:37.9511991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9512073Z outputs = self.roberta( 2025-09-07T07:40:37.9512507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9512598Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9513035Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9513125Z layer_outputs = layer_module( 2025-09-07T07:40:37.9513557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9513658Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9514090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9514195Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9514602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9514696Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9515178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9515335Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9515772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T07:40:37.9515881Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9515885Z 2025-09-07T07:40:37.9516022Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9516312Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9516386Z return mod(**inputs) 2025-09-07T07:40:37.9516824Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9516904Z outputs = self.roberta( 2025-09-07T07:40:37.9517335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9517427Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9517858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9517950Z layer_outputs = layer_module( 2025-09-07T07:40:37.9518295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9518393Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9518823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9518926Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9519333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9519425Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9519907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9520072Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9520505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T07:40:37.9520741Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T07:40:37.9521073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T07:40:37.9521157Z return self.act(input) 2025-09-07T07:40:37.9521162Z 2025-09-07T07:40:37.9521297Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9521587Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9521663Z return mod(**inputs) 2025-09-07T07:40:37.9522103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9522270Z outputs = self.roberta( 2025-09-07T07:40:37.9522702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9522796Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9523232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9523316Z layer_outputs = layer_module( 2025-09-07T07:40:37.9523663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9523759Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9524196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9524300Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9524707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9524803Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9525283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T07:40:37.9525464Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T07:40:37.9525893Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T07:40:37.9525999Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9526004Z 2025-09-07T07:40:37.9526139Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9526423Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9526504Z return mod(**inputs) 2025-09-07T07:40:37.9526944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9527030Z outputs = self.roberta( 2025-09-07T07:40:37.9527464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9527556Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9527983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9528067Z layer_outputs = layer_module( 2025-09-07T07:40:37.9528413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9528509Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9528954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9529066Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9529459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9529613Z return func(*args, **kwargs) 2025-09-07T07:40:37.9530071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9530159Z self_outputs = self.self( 2025-09-07T07:40:37.9530544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9530627Z return func(*args, **kwargs) 2025-09-07T07:40:37.9531068Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T07:40:37.9531373Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T07:40:37.9531429Z 2025-09-07T07:40:37.9531572Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9531860Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9531942Z return mod(**inputs) 2025-09-07T07:40:37.9532397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9532479Z outputs = self.roberta( 2025-09-07T07:40:37.9532917Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9533005Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9533440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9533526Z layer_outputs = layer_module( 2025-09-07T07:40:37.9533872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9533971Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9534405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9534512Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9534888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9534975Z return func(*args, **kwargs) 2025-09-07T07:40:37.9535406Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9535491Z self_outputs = self.self( 2025-09-07T07:40:37.9535873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9535960Z return func(*args, **kwargs) 2025-09-07T07:40:37.9536393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T07:40:37.9536478Z self.key(current_states) 2025-09-07T07:40:37.9536483Z 2025-09-07T07:40:37.9536619Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9536910Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9536985Z return mod(**inputs) 2025-09-07T07:40:37.9537431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9537512Z outputs = self.roberta( 2025-09-07T07:40:37.9537952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9538053Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9538495Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9538585Z layer_outputs = layer_module( 2025-09-07T07:40:37.9539005Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9539111Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9539542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9539647Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9540027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9540111Z return func(*args, **kwargs) 2025-09-07T07:40:37.9540544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9540678Z self_outputs = self.self( 2025-09-07T07:40:37.9541058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9541144Z return func(*args, **kwargs) 2025-09-07T07:40:37.9541578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T07:40:37.9541666Z self.value(current_states) 2025-09-07T07:40:37.9541671Z 2025-09-07T07:40:37.9541769Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.9541907Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9542197Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9542276Z return mod(**inputs) 2025-09-07T07:40:37.9542720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9542806Z outputs = self.roberta( 2025-09-07T07:40:37.9543246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9543336Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9543770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9543861Z layer_outputs = layer_module( 2025-09-07T07:40:37.9544202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9544301Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9544728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9544839Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9545214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9545300Z return func(*args, **kwargs) 2025-09-07T07:40:37.9545734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9545817Z self_outputs = self.self( 2025-09-07T07:40:37.9546195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9546277Z return func(*args, **kwargs) 2025-09-07T07:40:37.9546711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T07:40:37.9546889Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T07:40:37.9546898Z 2025-09-07T07:40:37.9547034Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9547326Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9547403Z return mod(**inputs) 2025-09-07T07:40:37.9547929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9548020Z outputs = self.roberta( 2025-09-07T07:40:37.9548450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9548543Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9548976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9549065Z layer_outputs = layer_module( 2025-09-07T07:40:37.9549406Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9549589Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9550025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9550129Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9550506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9550590Z return func(*args, **kwargs) 2025-09-07T07:40:37.9551019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T07:40:37.9551189Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T07:40:37.9551617Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T07:40:37.9551727Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9551732Z 2025-09-07T07:40:37.9551866Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9552159Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9552236Z return mod(**inputs) 2025-09-07T07:40:37.9552673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9552756Z outputs = self.roberta( 2025-09-07T07:40:37.9553190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9553281Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9553711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9553800Z layer_outputs = layer_module( 2025-09-07T07:40:37.9554142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9554238Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9554672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9554777Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9555182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9555276Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9555753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9555915Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9556349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T07:40:37.9556454Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9556458Z 2025-09-07T07:40:37.9556693Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9556986Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9557065Z return mod(**inputs) 2025-09-07T07:40:37.9557503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9557588Z outputs = self.roberta( 2025-09-07T07:40:37.9558022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9558116Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9558627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9558713Z layer_outputs = layer_module( 2025-09-07T07:40:37.9559063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9559159Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9559591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9559696Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9560098Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9560193Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9560668Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9560832Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9561269Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T07:40:37.9561415Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T07:40:37.9561742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T07:40:37.9561822Z return self.act(input) 2025-09-07T07:40:37.9561826Z 2025-09-07T07:40:37.9561962Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9562246Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9562325Z return mod(**inputs) 2025-09-07T07:40:37.9562762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9562846Z outputs = self.roberta( 2025-09-07T07:40:37.9563282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9563376Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9563810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9563895Z layer_outputs = layer_module( 2025-09-07T07:40:37.9564243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9564340Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9564769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9564878Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9565371Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9565469Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9566132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T07:40:37.9566322Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T07:40:37.9566756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T07:40:37.9566858Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9566862Z 2025-09-07T07:40:37.9567002Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9567289Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9567365Z return mod(**inputs) 2025-09-07T07:40:37.9567904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9567985Z outputs = self.roberta( 2025-09-07T07:40:37.9568424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9568513Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9568948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9569034Z layer_outputs = layer_module( 2025-09-07T07:40:37.9569372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9569471Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9569900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9570012Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9570392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9570481Z return func(*args, **kwargs) 2025-09-07T07:40:37.9570911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9570995Z self_outputs = self.self( 2025-09-07T07:40:37.9571378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9571461Z return func(*args, **kwargs) 2025-09-07T07:40:37.9571893Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T07:40:37.9572197Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T07:40:37.9572207Z 2025-09-07T07:40:37.9572342Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9572634Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9572709Z return mod(**inputs) 2025-09-07T07:40:37.9573150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9573231Z outputs = self.roberta( 2025-09-07T07:40:37.9573667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9573754Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9574184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9574281Z layer_outputs = layer_module( 2025-09-07T07:40:37.9574628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9574724Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9575235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9575345Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9575725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9575807Z return func(*args, **kwargs) 2025-09-07T07:40:37.9576242Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9576332Z self_outputs = self.self( 2025-09-07T07:40:37.9576709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9576860Z return func(*args, **kwargs) 2025-09-07T07:40:37.9577296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T07:40:37.9577377Z self.key(current_states) 2025-09-07T07:40:37.9577382Z 2025-09-07T07:40:37.9577525Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9577815Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9577896Z return mod(**inputs) 2025-09-07T07:40:37.9578335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9578415Z outputs = self.roberta( 2025-09-07T07:40:37.9578852Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9578947Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9579382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9579474Z layer_outputs = layer_module( 2025-09-07T07:40:37.9579821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9579918Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9580349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9580456Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9580831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9580917Z return func(*args, **kwargs) 2025-09-07T07:40:37.9581352Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9581435Z self_outputs = self.self( 2025-09-07T07:40:37.9581819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9581902Z return func(*args, **kwargs) 2025-09-07T07:40:37.9582337Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T07:40:37.9582421Z self.value(current_states) 2025-09-07T07:40:37.9582426Z 2025-09-07T07:40:37.9582525Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.9582661Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9582946Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9583027Z return mod(**inputs) 2025-09-07T07:40:37.9583472Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9583557Z outputs = self.roberta( 2025-09-07T07:40:37.9584086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9584183Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9584620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9584707Z layer_outputs = layer_module( 2025-09-07T07:40:37.9585052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9585148Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9585578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9585757Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9586132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9586222Z return func(*args, **kwargs) 2025-09-07T07:40:37.9586652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9586739Z self_outputs = self.self( 2025-09-07T07:40:37.9587115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9587197Z return func(*args, **kwargs) 2025-09-07T07:40:37.9587635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T07:40:37.9587810Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T07:40:37.9587819Z 2025-09-07T07:40:37.9587958Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9588245Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9588325Z return mod(**inputs) 2025-09-07T07:40:37.9588769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9588851Z outputs = self.roberta( 2025-09-07T07:40:37.9589286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9589377Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9589813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9589901Z layer_outputs = layer_module( 2025-09-07T07:40:37.9590248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9590349Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9590785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9590897Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9591272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9591355Z return func(*args, **kwargs) 2025-09-07T07:40:37.9591790Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T07:40:37.9591958Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T07:40:37.9592395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T07:40:37.9592505Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9592509Z 2025-09-07T07:40:37.9592651Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9593055Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9593138Z return mod(**inputs) 2025-09-07T07:40:37.9593585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9593666Z outputs = self.roberta( 2025-09-07T07:40:37.9594099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9594188Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9594617Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9594780Z layer_outputs = layer_module( 2025-09-07T07:40:37.9595123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9595226Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9595655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9595758Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9596163Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9596255Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9596736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9596895Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9597333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T07:40:37.9597438Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9597445Z 2025-09-07T07:40:37.9597579Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9597866Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9597942Z return mod(**inputs) 2025-09-07T07:40:37.9598384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9598465Z outputs = self.roberta( 2025-09-07T07:40:37.9598897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9598992Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9599428Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9599519Z layer_outputs = layer_module( 2025-09-07T07:40:37.9599863Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9599964Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9600394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9600499Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9600908Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9601001Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9601479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9601641Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9602163Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T07:40:37.9602314Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T07:40:37.9602641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T07:40:37.9602727Z return self.act(input) 2025-09-07T07:40:37.9602731Z 2025-09-07T07:40:37.9602865Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9603153Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9603230Z return mod(**inputs) 2025-09-07T07:40:37.9603670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9603823Z outputs = self.roberta( 2025-09-07T07:40:37.9604260Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9604351Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9604780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9604865Z layer_outputs = layer_module( 2025-09-07T07:40:37.9605212Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9605308Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9605741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9605848Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9606255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9606346Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9606822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T07:40:37.9607004Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T07:40:37.9607437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T07:40:37.9607548Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9607552Z 2025-09-07T07:40:37.9607688Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9607977Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9608057Z return mod(**inputs) 2025-09-07T07:40:37.9608496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9608581Z outputs = self.roberta( 2025-09-07T07:40:37.9609014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9609106Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9609536Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9609621Z layer_outputs = layer_module( 2025-09-07T07:40:37.9609966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9610060Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9610495Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9610605Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9611075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9611167Z return func(*args, **kwargs) 2025-09-07T07:40:37.9611599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9611688Z self_outputs = self.self( 2025-09-07T07:40:37.9612066Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9612153Z return func(*args, **kwargs) 2025-09-07T07:40:37.9612584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T07:40:37.9612965Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T07:40:37.9612970Z 2025-09-07T07:40:37.9613109Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9613401Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9613481Z return mod(**inputs) 2025-09-07T07:40:37.9613919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9614000Z outputs = self.roberta( 2025-09-07T07:40:37.9614438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9614526Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9614964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9615056Z layer_outputs = layer_module( 2025-09-07T07:40:37.9615402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9615505Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9615937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9616045Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9616420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9616506Z return func(*args, **kwargs) 2025-09-07T07:40:37.9616935Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9617019Z self_outputs = self.self( 2025-09-07T07:40:37.9617404Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9617486Z return func(*args, **kwargs) 2025-09-07T07:40:37.9617921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T07:40:37.9618004Z self.key(current_states) 2025-09-07T07:40:37.9618008Z 2025-09-07T07:40:37.9618147Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9618434Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9618509Z return mod(**inputs) 2025-09-07T07:40:37.9618954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9619035Z outputs = self.roberta( 2025-09-07T07:40:37.9619473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9619564Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9620081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9620176Z layer_outputs = layer_module( 2025-09-07T07:40:37.9620527Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9620625Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9621057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9621162Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9621542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9621698Z return func(*args, **kwargs) 2025-09-07T07:40:37.9622133Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9622216Z self_outputs = self.self( 2025-09-07T07:40:37.9622600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9622683Z return func(*args, **kwargs) 2025-09-07T07:40:37.9623114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T07:40:37.9623200Z self.value(current_states) 2025-09-07T07:40:37.9623204Z 2025-09-07T07:40:37.9623302Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.9623439Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9623727Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9623808Z return mod(**inputs) 2025-09-07T07:40:37.9624251Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9624332Z outputs = self.roberta( 2025-09-07T07:40:37.9624771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9624861Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9625287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9625376Z layer_outputs = layer_module( 2025-09-07T07:40:37.9625716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9625815Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9626244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9626354Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9626733Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9626816Z return func(*args, **kwargs) 2025-09-07T07:40:37.9627252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9627335Z self_outputs = self.self( 2025-09-07T07:40:37.9627711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9627794Z return func(*args, **kwargs) 2025-09-07T07:40:37.9628221Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T07:40:37.9628404Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T07:40:37.9628409Z 2025-09-07T07:40:37.9628542Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9628963Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9629044Z return mod(**inputs) 2025-09-07T07:40:37.9629486Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9629568Z outputs = self.roberta( 2025-09-07T07:40:37.9629999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9630092Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9630523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9630610Z layer_outputs = layer_module( 2025-09-07T07:40:37.9631018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9631113Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9631550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9631654Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9632031Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9632112Z return func(*args, **kwargs) 2025-09-07T07:40:37.9632546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T07:40:37.9632712Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T07:40:37.9633142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T07:40:37.9633254Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9633259Z 2025-09-07T07:40:37.9633398Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9633691Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9633766Z return mod(**inputs) 2025-09-07T07:40:37.9634204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9634293Z outputs = self.roberta( 2025-09-07T07:40:37.9634723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9634817Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9635248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9635343Z layer_outputs = layer_module( 2025-09-07T07:40:37.9635684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9635783Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9636216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9636319Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9636727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9636816Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9637293Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9637460Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9637891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T07:40:37.9638060Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9638065Z 2025-09-07T07:40:37.9638206Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9638498Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9638574Z return mod(**inputs) 2025-09-07T07:40:37.9639012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9639098Z outputs = self.roberta( 2025-09-07T07:40:37.9639529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9639669Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9640100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9640185Z layer_outputs = layer_module( 2025-09-07T07:40:37.9640536Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9640635Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9641070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9641175Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9641584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9641676Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9642152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9642319Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9642755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T07:40:37.9642900Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T07:40:37.9643227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T07:40:37.9643309Z return self.act(input) 2025-09-07T07:40:37.9643314Z 2025-09-07T07:40:37.9643450Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9643736Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9643817Z return mod(**inputs) 2025-09-07T07:40:37.9644256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9644343Z outputs = self.roberta( 2025-09-07T07:40:37.9644776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9644865Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9645300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9645387Z layer_outputs = layer_module( 2025-09-07T07:40:37.9645731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9645827Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9646257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9646369Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9646773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9646958Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9647445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T07:40:37.9647625Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T07:40:37.9648058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T07:40:37.9648160Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9648165Z 2025-09-07T07:40:37.9648303Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9648589Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9648730Z return mod(**inputs) 2025-09-07T07:40:37.9649173Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9649260Z outputs = self.roberta( 2025-09-07T07:40:37.9649697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9649785Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9650216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9650304Z layer_outputs = layer_module( 2025-09-07T07:40:37.9650646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9650746Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9651180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9651288Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9651667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9651753Z return func(*args, **kwargs) 2025-09-07T07:40:37.9652183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9652267Z self_outputs = self.self( 2025-09-07T07:40:37.9652647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9652729Z return func(*args, **kwargs) 2025-09-07T07:40:37.9653163Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T07:40:37.9653471Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T07:40:37.9653476Z 2025-09-07T07:40:37.9653615Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9653903Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9653980Z return mod(**inputs) 2025-09-07T07:40:37.9654421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9654503Z outputs = self.roberta( 2025-09-07T07:40:37.9654937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9655027Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9655457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9655555Z layer_outputs = layer_module( 2025-09-07T07:40:37.9655897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9656055Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9656491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9656596Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9656975Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9657057Z return func(*args, **kwargs) 2025-09-07T07:40:37.9657493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9657578Z self_outputs = self.self( 2025-09-07T07:40:37.9658045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9658128Z return func(*args, **kwargs) 2025-09-07T07:40:37.9658563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T07:40:37.9658650Z self.key(current_states) 2025-09-07T07:40:37.9658654Z 2025-09-07T07:40:37.9658789Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9659079Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9659154Z return mod(**inputs) 2025-09-07T07:40:37.9659593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9659678Z outputs = self.roberta( 2025-09-07T07:40:37.9660107Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9660205Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9660642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9660730Z layer_outputs = layer_module( 2025-09-07T07:40:37.9661085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9661184Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9661616Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9661724Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9662101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9662193Z return func(*args, **kwargs) 2025-09-07T07:40:37.9662626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9662710Z self_outputs = self.self( 2025-09-07T07:40:37.9663093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9663176Z return func(*args, **kwargs) 2025-09-07T07:40:37.9663611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T07:40:37.9663698Z self.value(current_states) 2025-09-07T07:40:37.9663703Z 2025-09-07T07:40:37.9663803Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.9663939Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9664226Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9664311Z return mod(**inputs) 2025-09-07T07:40:37.9664751Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9664909Z outputs = self.roberta( 2025-09-07T07:40:37.9665422Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9665515Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9665952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9666037Z layer_outputs = layer_module( 2025-09-07T07:40:37.9666384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9666481Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9666912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9667176Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9667558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9667646Z return func(*args, **kwargs) 2025-09-07T07:40:37.9668078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9668163Z self_outputs = self.self( 2025-09-07T07:40:37.9668538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9668621Z return func(*args, **kwargs) 2025-09-07T07:40:37.9669055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T07:40:37.9669237Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T07:40:37.9669241Z 2025-09-07T07:40:37.9669379Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9669670Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9669746Z return mod(**inputs) 2025-09-07T07:40:37.9670186Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9670268Z outputs = self.roberta( 2025-09-07T07:40:37.9670703Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9670792Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9671226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9671316Z layer_outputs = layer_module( 2025-09-07T07:40:37.9671658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9671758Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9672191Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9672301Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9672674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9672758Z return func(*args, **kwargs) 2025-09-07T07:40:37.9673192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T07:40:37.9673360Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T07:40:37.9673802Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T07:40:37.9673907Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9673911Z 2025-09-07T07:40:37.9674187Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9674488Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9674563Z return mod(**inputs) 2025-09-07T07:40:37.9675008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9675091Z outputs = self.roberta( 2025-09-07T07:40:37.9675528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9675617Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9676048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9676217Z layer_outputs = layer_module( 2025-09-07T07:40:37.9676563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9676663Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9677095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9677202Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9677604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9677696Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9678173Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9678338Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9678772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T07:40:37.9678879Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9678884Z 2025-09-07T07:40:37.9679018Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9679306Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9679382Z return mod(**inputs) 2025-09-07T07:40:37.9679821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9679905Z outputs = self.roberta( 2025-09-07T07:40:37.9680339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9680432Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9680865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9680960Z layer_outputs = layer_module( 2025-09-07T07:40:37.9681301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9681402Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9681835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9681939Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9682344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9682436Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9682919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9683078Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9683601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T07:40:37.9683751Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T07:40:37.9684077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T07:40:37.9684161Z return self.act(input) 2025-09-07T07:40:37.9684166Z 2025-09-07T07:40:37.9684303Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9684594Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9684670Z return mod(**inputs) 2025-09-07T07:40:37.9685213Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9685301Z outputs = self.roberta( 2025-09-07T07:40:37.9685750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9685843Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9686287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9686374Z layer_outputs = layer_module( 2025-09-07T07:40:37.9686732Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9686834Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9687281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9687391Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9687809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9687907Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9688400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T07:40:37.9688585Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T07:40:37.9689030Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T07:40:37.9689140Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9689145Z 2025-09-07T07:40:37.9689283Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9689584Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9689667Z return mod(**inputs) 2025-09-07T07:40:37.9690126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9690211Z outputs = self.roberta( 2025-09-07T07:40:37.9690656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9690750Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9691200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9691288Z layer_outputs = layer_module( 2025-09-07T07:40:37.9691646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9691747Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9692203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9692313Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9692824Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9692943Z return func(*args, **kwargs) 2025-09-07T07:40:37.9693388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9693478Z self_outputs = self.self( 2025-09-07T07:40:37.9693867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9693954Z return func(*args, **kwargs) 2025-09-07T07:40:37.9694401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T07:40:37.9694807Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T07:40:37.9694812Z 2025-09-07T07:40:37.9694957Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9695245Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9695328Z return mod(**inputs) 2025-09-07T07:40:37.9695766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9695846Z outputs = self.roberta( 2025-09-07T07:40:37.9696282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9696371Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9696807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9696897Z layer_outputs = layer_module( 2025-09-07T07:40:37.9697246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9697344Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9697775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9697885Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9698258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9698342Z return func(*args, **kwargs) 2025-09-07T07:40:37.9698771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9698860Z self_outputs = self.self( 2025-09-07T07:40:37.9699238Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9699321Z return func(*args, **kwargs) 2025-09-07T07:40:37.9699762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T07:40:37.9699845Z self.key(current_states) 2025-09-07T07:40:37.9699850Z 2025-09-07T07:40:37.9699985Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9700273Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9700348Z return mod(**inputs) 2025-09-07T07:40:37.9700788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9700870Z outputs = self.roberta( 2025-09-07T07:40:37.9701309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9701398Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9701922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9702015Z layer_outputs = layer_module( 2025-09-07T07:40:37.9702358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9702458Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9702889Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9702994Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9703369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9703522Z return func(*args, **kwargs) 2025-09-07T07:40:37.9703957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9704046Z self_outputs = self.self( 2025-09-07T07:40:37.9704423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9704505Z return func(*args, **kwargs) 2025-09-07T07:40:37.9704933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T07:40:37.9705021Z self.value(current_states) 2025-09-07T07:40:37.9705025Z 2025-09-07T07:40:37.9705121Z cudagraph partition due to non gpu ops 2025-09-07T07:40:37.9705262Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9705546Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9705626Z return mod(**inputs) 2025-09-07T07:40:37.9706068Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9706152Z outputs = self.roberta( 2025-09-07T07:40:37.9706590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9706680Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9707114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9707200Z layer_outputs = layer_module( 2025-09-07T07:40:37.9707542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9707641Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9708074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9708182Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9708558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9708643Z return func(*args, **kwargs) 2025-09-07T07:40:37.9709077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T07:40:37.9709161Z self_outputs = self.self( 2025-09-07T07:40:37.9709539Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9709621Z return func(*args, **kwargs) 2025-09-07T07:40:37.9710048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T07:40:37.9710233Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T07:40:37.9710238Z 2025-09-07T07:40:37.9710372Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9710719Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9710798Z return mod(**inputs) 2025-09-07T07:40:37.9711242Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9711322Z outputs = self.roberta( 2025-09-07T07:40:37.9711753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9711846Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9712277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9712430Z layer_outputs = layer_module( 2025-09-07T07:40:37.9712770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9712870Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9713306Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T07:40:37.9713411Z self_attention_outputs = self.attention( 2025-09-07T07:40:37.9713789Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T07:40:37.9713873Z return func(*args, **kwargs) 2025-09-07T07:40:37.9714305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T07:40:37.9714469Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T07:40:37.9714907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T07:40:37.9715015Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9715023Z 2025-09-07T07:40:37.9715160Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9715451Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9715527Z return mod(**inputs) 2025-09-07T07:40:37.9715963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9716049Z outputs = self.roberta( 2025-09-07T07:40:37.9716482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9716574Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9717006Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9717096Z layer_outputs = layer_module( 2025-09-07T07:40:37.9717444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9717542Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9717979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9718083Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9718497Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9718591Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9719086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9719258Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9719780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T07:40:37.9719901Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9719906Z 2025-09-07T07:40:37.9720044Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9720343Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9720418Z return mod(**inputs) 2025-09-07T07:40:37.9720864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9720949Z outputs = self.roberta( 2025-09-07T07:40:37.9721381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9721539Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9721974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9722060Z layer_outputs = layer_module( 2025-09-07T07:40:37.9722410Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9722506Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9722939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9723045Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9723452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9723550Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9724027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T07:40:37.9724191Z intermediate_output = self.intermediate(attention_output) 2025-09-07T07:40:37.9724624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T07:40:37.9724770Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T07:40:37.9725096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T07:40:37.9725176Z return self.act(input) 2025-09-07T07:40:37.9725185Z 2025-09-07T07:40:37.9725319Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9725605Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9725689Z return mod(**inputs) 2025-09-07T07:40:37.9726131Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T07:40:37.9726216Z outputs = self.roberta( 2025-09-07T07:40:37.9726671Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T07:40:37.9726774Z encoder_outputs = self.encoder( 2025-09-07T07:40:37.9727214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T07:40:37.9727300Z layer_outputs = layer_module( 2025-09-07T07:40:37.9727644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:40:37.9727740Z return super().__call__(*args, **kwargs) 2025-09-07T07:40:37.9728173Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T07:40:37.9728285Z layer_output = apply_chunking_to_forward( 2025-09-07T07:40:37.9728783Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T07:40:37.9728884Z return forward_fn(*input_tensors) 2025-09-07T07:40:37.9729362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T07:40:37.9729544Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T07:40:37.9729978Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T07:40:37.9730081Z hidden_states = self.dense(hidden_states) 2025-09-07T07:40:37.9730086Z 2025-09-07T07:40:37.9730226Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9730586Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9730667Z return mod(**inputs) 2025-09-07T07:40:37.9731113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1108, in forward 2025-09-07T07:40:37.9731244Z prediction_scores = self.lm_head(sequence_output) 2025-09-07T07:40:37.9731688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1143, in forward 2025-09-07T07:40:37.9731770Z x = self.dense(features) 2025-09-07T07:40:37.9731775Z 2025-09-07T07:40:37.9731912Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:40:37.9732201Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:40:37.9732280Z return mod(**inputs) 2025-09-07T07:40:37.9732719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1108, in forward 2025-09-07T07:40:37.9732849Z prediction_scores = self.lm_head(sequence_output) 2025-09-07T07:40:37.9733295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1148, in forward 2025-09-07T07:40:37.9733372Z x = self.decoder(x) 2025-09-07T07:40:37.9733377Z 2025-09-07T07:40:45.3344378Z pass 2025-09-07T07:40:45.3344859Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:40:47.9542174Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:40:47.9543455Z import pynvml # type: ignore[import] 2025-09-07T07:40:50.5542118Z 2025-09-07T07:40:53.9321560Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:40:53.9321835Z loading model: 0it [00:03, ?it/s] 2025-09-07T07:40:53.9408717Z cpu eval hf_T5 2025-09-07T07:40:55.6370718Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:40:56.0635050Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:40:56.4850840Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:41:14.9601527Z cudagraph partition due to non gpu ops 2025-09-07T07:41:14.9601910Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9602456Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9602926Z return mod(**inputs) 2025-09-07T07:41:14.9603464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:14.9604075Z decoder_outputs = self.decoder( 2025-09-07T07:41:14.9604623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9605448Z layer_outputs = layer_module( 2025-09-07T07:41:14.9605994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9606543Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9607106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9607677Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9608235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9608812Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9609377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 546, in forward 2025-09-07T07:41:14.9610129Z position_bias = position_bias + causal_mask 2025-09-07T07:41:14.9610335Z 2025-09-07T07:41:14.9610607Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9611133Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9611598Z return mod(**inputs) 2025-09-07T07:41:14.9612107Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:14.9612663Z decoder_outputs = self.decoder( 2025-09-07T07:41:14.9613206Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9613762Z layer_outputs = layer_module( 2025-09-07T07:41:14.9614272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9614817Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9615378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9615946Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9616505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:41:14.9617107Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:14.9617704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:14.9618259Z return self.weight * hidden_states 2025-09-07T07:41:14.9618444Z 2025-09-07T07:41:14.9618579Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9619108Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9619574Z return mod(**inputs) 2025-09-07T07:41:14.9620090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:14.9620650Z decoder_outputs = self.decoder( 2025-09-07T07:41:14.9621194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9621750Z layer_outputs = layer_module( 2025-09-07T07:41:14.9622260Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9622803Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9623361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9623927Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9624487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9625058Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9625618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:14.9626300Z query_states = self.q(hidden_states) 2025-09-07T07:41:14.9626502Z 2025-09-07T07:41:14.9626641Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9627164Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9627631Z return mod(**inputs) 2025-09-07T07:41:14.9628142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:14.9628697Z decoder_outputs = self.decoder( 2025-09-07T07:41:14.9629242Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9629897Z layer_outputs = layer_module( 2025-09-07T07:41:14.9630414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9630953Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9631513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9632079Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9632644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9633212Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9633771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:14.9634323Z key_states = self.k(current_states) 2025-09-07T07:41:14.9634514Z 2025-09-07T07:41:14.9634656Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9635180Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9635648Z return mod(**inputs) 2025-09-07T07:41:14.9636161Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:14.9636720Z decoder_outputs = self.decoder( 2025-09-07T07:41:14.9637265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9637823Z layer_outputs = layer_module( 2025-09-07T07:41:14.9638330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9638874Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9639427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9639995Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9640557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9641125Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9641695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:14.9642334Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:14.9642600Z 2025-09-07T07:41:14.9642739Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9643263Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9643722Z return mod(**inputs) 2025-09-07T07:41:14.9644232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:14.9644793Z decoder_outputs = self.decoder( 2025-09-07T07:41:14.9645345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9645895Z layer_outputs = layer_module( 2025-09-07T07:41:14.9646531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9647086Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9647644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9648208Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9648759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9649326Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9649886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:14.9650612Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:14.9650885Z 2025-09-07T07:41:14.9651029Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9651546Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9652012Z return mod(**inputs) 2025-09-07T07:41:14.9652522Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:14.9653080Z decoder_outputs = self.decoder( 2025-09-07T07:41:14.9653624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9654178Z layer_outputs = layer_module( 2025-09-07T07:41:14.9654691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9655234Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9655800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9656362Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9656931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9657499Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9658064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:14.9658695Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:14.9658964Z 2025-09-07T07:41:14.9659098Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9659618Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9660084Z return mod(**inputs) 2025-09-07T07:41:14.9660606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:14.9661166Z decoder_outputs = self.decoder( 2025-09-07T07:41:14.9661710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9662268Z layer_outputs = layer_module( 2025-09-07T07:41:14.9662780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9663322Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9663875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9664443Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9664998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9665673Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9666236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:14.9667085Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:14.9667414Z 2025-09-07T07:41:14.9667553Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9668075Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9668561Z return mod(**inputs) 2025-09-07T07:41:14.9669071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:14.9669623Z decoder_outputs = self.decoder( 2025-09-07T07:41:14.9670163Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9670806Z layer_outputs = layer_module( 2025-09-07T07:41:14.9671318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9671849Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9672409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9672978Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9673535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9674099Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9674660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:14.9675220Z value_states = self.v(current_states) 2025-09-07T07:41:14.9675414Z 2025-09-07T07:41:14.9675550Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9676073Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9676539Z return mod(**inputs) 2025-09-07T07:41:14.9677051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:14.9677614Z decoder_outputs = self.decoder( 2025-09-07T07:41:14.9678157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9678711Z layer_outputs = layer_module( 2025-09-07T07:41:14.9679212Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9679757Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9680310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9693765Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9694485Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9695071Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9695649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:14.9696262Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:14.9696501Z 2025-09-07T07:41:14.9696643Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9697165Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9697626Z return mod(**inputs) 2025-09-07T07:41:14.9698140Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:14.9698697Z decoder_outputs = self.decoder( 2025-09-07T07:41:14.9699249Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9699806Z layer_outputs = layer_module( 2025-09-07T07:41:14.9700536Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9701088Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9701652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9702214Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9702771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9703337Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9703899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:14.9704578Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:14.9704819Z 2025-09-07T07:41:14.9704957Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9705481Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9705945Z return mod(**inputs) 2025-09-07T07:41:14.9706447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:14.9707007Z decoder_outputs = self.decoder( 2025-09-07T07:41:14.9707548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9708101Z layer_outputs = layer_module( 2025-09-07T07:41:14.9708608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9709139Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9709703Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9710264Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9710832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9711397Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9711953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:14.9712553Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:14.9712789Z 2025-09-07T07:41:14.9712925Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9713444Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9713904Z return mod(**inputs) 2025-09-07T07:41:14.9714408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:14.9714967Z decoder_outputs = self.decoder( 2025-09-07T07:41:14.9715513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9716064Z layer_outputs = layer_module( 2025-09-07T07:41:14.9716567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9717103Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9717659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9718218Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9718776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9719336Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9719902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:14.9720507Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:14.9720849Z 2025-09-07T07:41:14.9720997Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9721513Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9721976Z return mod(**inputs) 2025-09-07T07:41:14.9722482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:14.9723033Z decoder_outputs = self.decoder( 2025-09-07T07:41:14.9723573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9724116Z layer_outputs = layer_module( 2025-09-07T07:41:14.9724714Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9725247Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9725805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9726367Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9726915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9727477Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9728036Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:14.9728586Z attn_output = self.o(attn_output) 2025-09-07T07:41:14.9728763Z 2025-09-07T07:41:14.9728896Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9729415Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9729879Z return mod(**inputs) 2025-09-07T07:41:14.9730384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:14.9730942Z decoder_outputs = self.decoder( 2025-09-07T07:41:14.9731476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9732027Z layer_outputs = layer_module( 2025-09-07T07:41:14.9732533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9733066Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9733614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:14.9734177Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:14.9734738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:14.9735305Z attention_output = self.EncDecAttention( 2025-09-07T07:41:14.9735868Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:14.9736416Z query_states = self.q(hidden_states) 2025-09-07T07:41:14.9736607Z 2025-09-07T07:41:14.9736742Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9737258Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9737717Z return mod(**inputs) 2025-09-07T07:41:14.9738225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9738773Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9739315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9739870Z layer_outputs = layer_module( 2025-09-07T07:41:14.9740376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9741028Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9741596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9742158Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9742711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9743274Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9743825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:14.9744379Z query_states = self.q(hidden_states) 2025-09-07T07:41:14.9744654Z 2025-09-07T07:41:14.9744791Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9745306Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9745765Z return mod(**inputs) 2025-09-07T07:41:14.9746272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9746825Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9747363Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9747910Z layer_outputs = layer_module( 2025-09-07T07:41:14.9748413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9748948Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9749499Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9750064Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9750620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9751181Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9751741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:14.9752294Z key_states = self.k(current_states) 2025-09-07T07:41:14.9752476Z 2025-09-07T07:41:14.9752616Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9753129Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9753586Z return mod(**inputs) 2025-09-07T07:41:14.9754092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9754653Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9755191Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9755738Z layer_outputs = layer_module( 2025-09-07T07:41:14.9756247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9756781Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9757333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9757893Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9758445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9759007Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9759565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:14.9760199Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:14.9760465Z 2025-09-07T07:41:14.9760602Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9761254Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9761741Z return mod(**inputs) 2025-09-07T07:41:14.9762244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9762797Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9763329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9763877Z layer_outputs = layer_module( 2025-09-07T07:41:14.9764385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9765008Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9765645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9766212Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9766769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9767327Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9767884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:14.9768507Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:14.9768773Z 2025-09-07T07:41:14.9768911Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9769422Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9769885Z return mod(**inputs) 2025-09-07T07:41:14.9770391Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9770944Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9771489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9772035Z layer_outputs = layer_module( 2025-09-07T07:41:14.9772535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9773067Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9773619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9774177Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9774726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9775293Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9775851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:14.9776482Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:14.9776750Z 2025-09-07T07:41:14.9776888Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9777400Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9777866Z return mod(**inputs) 2025-09-07T07:41:14.9778372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9778925Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9779461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9780014Z layer_outputs = layer_module( 2025-09-07T07:41:14.9780518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9781051Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9781796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9782387Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9782944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9783506Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9784062Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:14.9784730Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:14.9785167Z 2025-09-07T07:41:14.9785305Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9785822Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9786282Z return mod(**inputs) 2025-09-07T07:41:14.9786796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9787351Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9787886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9788435Z layer_outputs = layer_module( 2025-09-07T07:41:14.9788940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9789472Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9790019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9790584Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9791141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9791709Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9792267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:14.9792932Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:14.9793241Z 2025-09-07T07:41:14.9793376Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9793891Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9794347Z return mod(**inputs) 2025-09-07T07:41:14.9794853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9795407Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9795941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9796489Z layer_outputs = layer_module( 2025-09-07T07:41:14.9796995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9797526Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9798079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9798636Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9799190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9799751Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9800303Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:14.9800863Z value_states = self.v(current_states) 2025-09-07T07:41:14.9801052Z 2025-09-07T07:41:14.9801187Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9801815Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9802285Z return mod(**inputs) 2025-09-07T07:41:14.9802793Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9803345Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9803884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9804431Z layer_outputs = layer_module( 2025-09-07T07:41:14.9804928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9805549Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9806101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9806667Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9807221Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9807776Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9808335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:14.9808934Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:14.9809167Z 2025-09-07T07:41:14.9809306Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9809813Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9810275Z return mod(**inputs) 2025-09-07T07:41:14.9810781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9811335Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9811881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9812427Z layer_outputs = layer_module( 2025-09-07T07:41:14.9812931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9813465Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9814015Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9814572Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9815119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9815690Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9816244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:14.9817671Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:14.9817908Z 2025-09-07T07:41:14.9818047Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9818560Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9819020Z return mod(**inputs) 2025-09-07T07:41:14.9819528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9820082Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9820617Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9821171Z layer_outputs = layer_module( 2025-09-07T07:41:14.9821677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9822214Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9822891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9823457Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9824014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9824570Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9825123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:14.9825716Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:14.9825951Z 2025-09-07T07:41:14.9826086Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9826709Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9827169Z return mod(**inputs) 2025-09-07T07:41:14.9827680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9828228Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9828770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9829320Z layer_outputs = layer_module( 2025-09-07T07:41:14.9829826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9830361Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9830909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9831476Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9832032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9832595Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9833153Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:14.9833754Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:14.9833990Z 2025-09-07T07:41:14.9834125Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9834637Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9835097Z return mod(**inputs) 2025-09-07T07:41:14.9835599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9836154Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9836696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9837242Z layer_outputs = layer_module( 2025-09-07T07:41:14.9837752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9838280Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9838837Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9839397Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9839951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9840510Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9841069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:14.9841625Z attn_output = self.o(attn_output) 2025-09-07T07:41:14.9841802Z 2025-09-07T07:41:14.9841906Z cudagraph partition due to non gpu ops 2025-09-07T07:41:14.9842224Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9842847Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9843320Z return mod(**inputs) 2025-09-07T07:41:14.9843829Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9844384Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9844918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9845467Z layer_outputs = layer_module( 2025-09-07T07:41:14.9845973Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9846597Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9847147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:14.9847718Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:14.9848293Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:41:14.9848881Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:41:14.9849464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:14.9850014Z return self.weight * hidden_states 2025-09-07T07:41:14.9850197Z 2025-09-07T07:41:14.9850332Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9850846Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9851311Z return mod(**inputs) 2025-09-07T07:41:14.9851815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9852366Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9852909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9853459Z layer_outputs = layer_module( 2025-09-07T07:41:14.9853966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9854499Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9855048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:14.9855619Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:14.9856189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:14.9856808Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:14.9857419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:41:14.9857994Z hidden_states = self.wi(hidden_states) 2025-09-07T07:41:14.9858183Z 2025-09-07T07:41:14.9858323Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9858840Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9859300Z return mod(**inputs) 2025-09-07T07:41:14.9859811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9860369Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9860908Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9861461Z layer_outputs = layer_module( 2025-09-07T07:41:14.9861966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9862500Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9863173Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:14.9863763Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:14.9864341Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:14.9864955Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:14.9865654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:41:14.9866218Z hidden_states = self.act(hidden_states) 2025-09-07T07:41:14.9866409Z 2025-09-07T07:41:14.9866704Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9867220Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9867682Z return mod(**inputs) 2025-09-07T07:41:14.9868193Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9868745Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9869282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9869829Z layer_outputs = layer_module( 2025-09-07T07:41:14.9870336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9870869Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9871422Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:14.9871994Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:14.9872565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:14.9873181Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:14.9873793Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:41:14.9874344Z hidden_states = self.wo(hidden_states) 2025-09-07T07:41:14.9874533Z 2025-09-07T07:41:14.9874631Z cudagraph partition due to non gpu ops 2025-09-07T07:41:14.9874948Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9875463Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9875924Z return mod(**inputs) 2025-09-07T07:41:14.9876426Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9876983Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9877523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9878077Z layer_outputs = layer_module( 2025-09-07T07:41:14.9878582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9879119Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9879673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9880233Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9880789Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:41:14.9881379Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:14.9881979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:14.9882533Z return self.weight * hidden_states 2025-09-07T07:41:14.9882715Z 2025-09-07T07:41:14.9883006Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9883534Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9883991Z return mod(**inputs) 2025-09-07T07:41:14.9884501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9885054Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9885594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9886137Z layer_outputs = layer_module( 2025-09-07T07:41:14.9886643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9887267Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9887821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9888382Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9888937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9889497Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9890060Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:14.9890611Z query_states = self.q(hidden_states) 2025-09-07T07:41:14.9890795Z 2025-09-07T07:41:14.9890935Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9891448Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9891911Z return mod(**inputs) 2025-09-07T07:41:14.9892415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9892969Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9893510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9894061Z layer_outputs = layer_module( 2025-09-07T07:41:14.9894566Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9895099Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9895649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9896203Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9896759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9897328Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9897885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:14.9898440Z key_states = self.k(current_states) 2025-09-07T07:41:14.9898623Z 2025-09-07T07:41:14.9898758Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9899274Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9899733Z return mod(**inputs) 2025-09-07T07:41:14.9900237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9900790Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9901329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9901886Z layer_outputs = layer_module( 2025-09-07T07:41:14.9902393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9902927Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9903590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9904163Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9904719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9905277Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9905833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:14.9906460Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:14.9906729Z 2025-09-07T07:41:14.9906974Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9907491Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9907951Z return mod(**inputs) 2025-09-07T07:41:14.9908460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9909015Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9909552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9910098Z layer_outputs = layer_module( 2025-09-07T07:41:14.9910601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9911129Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9911687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9912253Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9912805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9913369Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9913926Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:14.9914553Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:14.9914818Z 2025-09-07T07:41:14.9914956Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9915470Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9915927Z return mod(**inputs) 2025-09-07T07:41:14.9916432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9916990Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9917532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9918082Z layer_outputs = layer_module( 2025-09-07T07:41:14.9918588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9919123Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9919675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9920233Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9920789Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9921354Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9921911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:14.9922546Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:14.9922812Z 2025-09-07T07:41:14.9922950Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9923572Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9924039Z return mod(**inputs) 2025-09-07T07:41:14.9924544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9925094Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9925631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9926175Z layer_outputs = layer_module( 2025-09-07T07:41:14.9926681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9927313Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9927864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9928420Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9928979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9929540Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9930098Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:14.9930764Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:14.9931073Z 2025-09-07T07:41:14.9931206Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9931721Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9932184Z return mod(**inputs) 2025-09-07T07:41:14.9932689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9933242Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9933779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9934327Z layer_outputs = layer_module( 2025-09-07T07:41:14.9934833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9935365Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9935912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9936474Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9937031Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9937599Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9938156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:14.9938826Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:14.9939138Z 2025-09-07T07:41:14.9939272Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9939786Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9940243Z return mod(**inputs) 2025-09-07T07:41:14.9940747Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9941296Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9941833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9942388Z layer_outputs = layer_module( 2025-09-07T07:41:14.9942890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9943421Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9944094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9944665Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9945224Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9945786Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9946338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:14.9946890Z value_states = self.v(current_states) 2025-09-07T07:41:14.9947079Z 2025-09-07T07:41:14.9947302Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9947817Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9948274Z return mod(**inputs) 2025-09-07T07:41:14.9948778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9949330Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9949870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9950416Z layer_outputs = layer_module( 2025-09-07T07:41:14.9950914Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9951449Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9951997Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9952560Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9953114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9953674Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9954231Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:14.9954826Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:14.9955058Z 2025-09-07T07:41:14.9955196Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9955719Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9956177Z return mod(**inputs) 2025-09-07T07:41:14.9956680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9957230Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9957774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9958321Z layer_outputs = layer_module( 2025-09-07T07:41:14.9958830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9959368Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9959920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9960479Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9961033Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9961600Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9962161Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:14.9962763Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:14.9962997Z 2025-09-07T07:41:14.9963134Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9963756Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9964226Z return mod(**inputs) 2025-09-07T07:41:14.9964736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9965352Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9965894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9966443Z layer_outputs = layer_module( 2025-09-07T07:41:14.9966949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9967627Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9968178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9968733Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9969292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9969854Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9970410Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:14.9971005Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:14.9971244Z 2025-09-07T07:41:14.9971377Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9971890Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9972352Z return mod(**inputs) 2025-09-07T07:41:14.9972861Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9973410Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9973949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9974497Z layer_outputs = layer_module( 2025-09-07T07:41:14.9975002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9975534Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9976083Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9976644Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9977201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9977769Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9978322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:14.9978930Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:14.9979166Z 2025-09-07T07:41:14.9979301Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9979813Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9980274Z return mod(**inputs) 2025-09-07T07:41:14.9980777Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9981336Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9981874Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9982423Z layer_outputs = layer_module( 2025-09-07T07:41:14.9982935Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9983470Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9984211Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:14.9984781Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:14.9985338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:14.9985897Z attention_output = self.SelfAttention( 2025-09-07T07:41:14.9986456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:14.9987010Z attn_output = self.o(attn_output) 2025-09-07T07:41:14.9987189Z 2025-09-07T07:41:14.9987328Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9987933Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9988389Z return mod(**inputs) 2025-09-07T07:41:14.9988899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9989453Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9989995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9990542Z layer_outputs = layer_module( 2025-09-07T07:41:14.9991047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9991583Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:14.9992134Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:14.9992705Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:14.9993276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:41:14.9993862Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:41:14.9994447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:14.9994999Z return self.weight * hidden_states 2025-09-07T07:41:14.9995183Z 2025-09-07T07:41:14.9995321Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:14.9995830Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:14.9996290Z return mod(**inputs) 2025-09-07T07:41:14.9996797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:14.9997350Z encoder_outputs = self.encoder( 2025-09-07T07:41:14.9997888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:14.9998445Z layer_outputs = layer_module( 2025-09-07T07:41:14.9998957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:14.9999492Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0000043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0000611Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0001183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0001801Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0002415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:41:15.0002975Z hidden_states = self.wi(hidden_states) 2025-09-07T07:41:15.0003164Z 2025-09-07T07:41:15.0003300Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0003817Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0004387Z return mod(**inputs) 2025-09-07T07:41:15.0004900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0005452Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0005992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0006542Z layer_outputs = layer_module( 2025-09-07T07:41:15.0007048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0007581Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0008220Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0008793Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0009365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0009974Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0010585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:41:15.0011139Z hidden_states = self.act(hidden_states) 2025-09-07T07:41:15.0011333Z 2025-09-07T07:41:15.0011469Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0011985Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0012446Z return mod(**inputs) 2025-09-07T07:41:15.0012948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0013510Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0014057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0014608Z layer_outputs = layer_module( 2025-09-07T07:41:15.0015115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0015648Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0016200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0016772Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0017342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0017952Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0018567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:41:15.0019125Z hidden_states = self.wo(hidden_states) 2025-09-07T07:41:15.0019318Z 2025-09-07T07:41:15.0019421Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0019741Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0020252Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0020715Z return mod(**inputs) 2025-09-07T07:41:15.0021221Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0021774Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0022310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0022860Z layer_outputs = layer_module( 2025-09-07T07:41:15.0023370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0023908Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0024579Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0025147Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0025703Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:41:15.0026301Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0026897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0027447Z return self.weight * hidden_states 2025-09-07T07:41:15.0027629Z 2025-09-07T07:41:15.0027764Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0028363Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0028824Z return mod(**inputs) 2025-09-07T07:41:15.0029331Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0029886Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0030421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0030971Z layer_outputs = layer_module( 2025-09-07T07:41:15.0031477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0032013Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0032564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0033128Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0033684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0034247Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0034809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:15.0035363Z query_states = self.q(hidden_states) 2025-09-07T07:41:15.0035553Z 2025-09-07T07:41:15.0035687Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0036202Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0036662Z return mod(**inputs) 2025-09-07T07:41:15.0037169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0037720Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0038264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0038813Z layer_outputs = layer_module( 2025-09-07T07:41:15.0039325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0039857Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0040409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0040972Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0041526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0042090Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0042645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:15.0043201Z key_states = self.k(current_states) 2025-09-07T07:41:15.0043386Z 2025-09-07T07:41:15.0043520Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0044036Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0044613Z return mod(**inputs) 2025-09-07T07:41:15.0045127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0045680Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0046218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0046766Z layer_outputs = layer_module( 2025-09-07T07:41:15.0047267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0047801Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0048442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0049001Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0049561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0050123Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0050681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0051312Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0051578Z 2025-09-07T07:41:15.0051716Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0052234Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0052694Z return mod(**inputs) 2025-09-07T07:41:15.0053196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0053766Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0054308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0054854Z layer_outputs = layer_module( 2025-09-07T07:41:15.0055357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0055889Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0056442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0057002Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0057556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0058118Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0058684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0059313Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0059580Z 2025-09-07T07:41:15.0059722Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0060237Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0060699Z return mod(**inputs) 2025-09-07T07:41:15.0061204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0061759Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0062297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0062845Z layer_outputs = layer_module( 2025-09-07T07:41:15.0063352Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0063891Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0064587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0065151Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0065787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0066349Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0066906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0067532Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0067800Z 2025-09-07T07:41:15.0067936Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0068449Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0069067Z return mod(**inputs) 2025-09-07T07:41:15.0069572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0070123Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0070659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0071208Z layer_outputs = layer_module( 2025-09-07T07:41:15.0071714Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0072249Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0072881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0073511Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0074291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0089105Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0089720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0090404Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0090712Z 2025-09-07T07:41:15.0090852Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0091369Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0091834Z return mod(**inputs) 2025-09-07T07:41:15.0092346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0092904Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0093447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0094007Z layer_outputs = layer_module( 2025-09-07T07:41:15.0094514Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0095048Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0095600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0096159Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0096715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0097278Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0097839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0098518Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0098827Z 2025-09-07T07:41:15.0098966Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0099706Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0100179Z return mod(**inputs) 2025-09-07T07:41:15.0100686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0101242Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0101779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0102327Z layer_outputs = layer_module( 2025-09-07T07:41:15.0102832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0103367Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0104021Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0104579Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0105136Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0105699Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0106259Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:15.0106817Z value_states = self.v(current_states) 2025-09-07T07:41:15.0107004Z 2025-09-07T07:41:15.0107139Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0107652Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0108107Z return mod(**inputs) 2025-09-07T07:41:15.0108615Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0109174Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0109714Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0110263Z layer_outputs = layer_module( 2025-09-07T07:41:15.0110768Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0111304Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0111851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0112410Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0112964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0113524Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0114081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0114678Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0114914Z 2025-09-07T07:41:15.0115052Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0115566Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0116023Z return mod(**inputs) 2025-09-07T07:41:15.0116530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0117079Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0117620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0118167Z layer_outputs = layer_module( 2025-09-07T07:41:15.0118673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0119207Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0119873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0120444Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0121002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0121565Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0122116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0122712Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0122948Z 2025-09-07T07:41:15.0123081Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0123593Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0124140Z return mod(**inputs) 2025-09-07T07:41:15.0124641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0125195Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0125734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0126281Z layer_outputs = layer_module( 2025-09-07T07:41:15.0126781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0127312Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0127862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0128420Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0128978Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0129536Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0130095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0130692Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0130924Z 2025-09-07T07:41:15.0131061Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0131571Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0132021Z return mod(**inputs) 2025-09-07T07:41:15.0132528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0133079Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0133616Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0134166Z layer_outputs = layer_module( 2025-09-07T07:41:15.0134670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0135206Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0135762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0136321Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0136873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0137435Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0137989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:15.0138586Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:15.0138826Z 2025-09-07T07:41:15.0138964Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0139472Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0140042Z return mod(**inputs) 2025-09-07T07:41:15.0140561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0141116Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0141652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0142195Z layer_outputs = layer_module( 2025-09-07T07:41:15.0142699Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0143230Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0143779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0144420Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0144976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0145534Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0146092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:15.0146638Z attn_output = self.o(attn_output) 2025-09-07T07:41:15.0146819Z 2025-09-07T07:41:15.0146915Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0147231Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0147742Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0148195Z return mod(**inputs) 2025-09-07T07:41:15.0148696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0149249Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0149791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0150340Z layer_outputs = layer_module( 2025-09-07T07:41:15.0150840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0151371Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0151921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0152491Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0153063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:41:15.0153646Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0154235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0154784Z return self.weight * hidden_states 2025-09-07T07:41:15.0154964Z 2025-09-07T07:41:15.0155105Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0155621Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0156072Z return mod(**inputs) 2025-09-07T07:41:15.0156579Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0157129Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0157667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0158211Z layer_outputs = layer_module( 2025-09-07T07:41:15.0158717Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0159257Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0159913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0160493Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0161057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0161666Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0162276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:41:15.0162833Z hidden_states = self.wi(hidden_states) 2025-09-07T07:41:15.0163021Z 2025-09-07T07:41:15.0163161Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0163670Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0164229Z return mod(**inputs) 2025-09-07T07:41:15.0164731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0165374Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0165912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0166456Z layer_outputs = layer_module( 2025-09-07T07:41:15.0166961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0167493Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0168040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0168606Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0169180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0169792Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0170406Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:41:15.0170962Z hidden_states = self.act(hidden_states) 2025-09-07T07:41:15.0171152Z 2025-09-07T07:41:15.0171286Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0171801Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0172256Z return mod(**inputs) 2025-09-07T07:41:15.0172759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0173304Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0173841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0174392Z layer_outputs = layer_module( 2025-09-07T07:41:15.0174895Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0175429Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0175976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0176547Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0177115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0177726Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0178336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:41:15.0178886Z hidden_states = self.wo(hidden_states) 2025-09-07T07:41:15.0178895Z 2025-09-07T07:41:15.0178993Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0179128Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0179594Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0179709Z return mod(**inputs) 2025-09-07T07:41:15.0180081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0180173Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0180539Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0180628Z layer_outputs = layer_module( 2025-09-07T07:41:15.0180967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0181063Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0181536Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0181635Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0182005Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:41:15.0182140Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0182501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0182597Z return self.weight * hidden_states 2025-09-07T07:41:15.0182602Z 2025-09-07T07:41:15.0182738Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0183025Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0183101Z return mod(**inputs) 2025-09-07T07:41:15.0183475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0183561Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0183930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0184020Z layer_outputs = layer_module( 2025-09-07T07:41:15.0184361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0184459Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0184821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0184920Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0185295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0185404Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0185771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:15.0185868Z query_states = self.q(hidden_states) 2025-09-07T07:41:15.0185877Z 2025-09-07T07:41:15.0186012Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0186303Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0186379Z return mod(**inputs) 2025-09-07T07:41:15.0186746Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0186838Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0187205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0187291Z layer_outputs = layer_module( 2025-09-07T07:41:15.0187638Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0187733Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0188197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0188305Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0188670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0188773Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0189140Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:15.0189231Z key_states = self.k(current_states) 2025-09-07T07:41:15.0189236Z 2025-09-07T07:41:15.0189371Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0189660Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0189852Z return mod(**inputs) 2025-09-07T07:41:15.0190223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0190317Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0190683Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0190773Z layer_outputs = layer_module( 2025-09-07T07:41:15.0191115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0191213Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0191577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0191677Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0192046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0192148Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0192516Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0192686Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0192692Z 2025-09-07T07:41:15.0192828Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0193111Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0193186Z return mod(**inputs) 2025-09-07T07:41:15.0193557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0193646Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0194013Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0194102Z layer_outputs = layer_module( 2025-09-07T07:41:15.0194446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0194548Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0194910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0195014Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0195377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0195484Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0195846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0196016Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0196025Z 2025-09-07T07:41:15.0196165Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0196452Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0196625Z return mod(**inputs) 2025-09-07T07:41:15.0196998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0197088Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0197461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0197546Z layer_outputs = layer_module( 2025-09-07T07:41:15.0197891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0197987Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0198351Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0198524Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0198892Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0198998Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0199361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0199531Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0199536Z 2025-09-07T07:41:15.0199670Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0199954Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0200031Z return mod(**inputs) 2025-09-07T07:41:15.0200398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0200494Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0200859Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0200946Z layer_outputs = layer_module( 2025-09-07T07:41:15.0201292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0201388Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0201753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0201851Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0202214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0202316Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0202682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0202895Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0202900Z 2025-09-07T07:41:15.0203037Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0203325Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0203402Z return mod(**inputs) 2025-09-07T07:41:15.0203768Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0203860Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0204227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0204316Z layer_outputs = layer_module( 2025-09-07T07:41:15.0204657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0204762Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0205228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0205369Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0205735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0205839Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0206204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0206413Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0206418Z 2025-09-07T07:41:15.0206553Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0206905Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0206981Z return mod(**inputs) 2025-09-07T07:41:15.0207353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0207444Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0207812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0207901Z layer_outputs = layer_module( 2025-09-07T07:41:15.0208241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0208341Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0208704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0208807Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0209174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0209276Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0209644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:15.0209740Z value_states = self.v(current_states) 2025-09-07T07:41:15.0209745Z 2025-09-07T07:41:15.0209882Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0210169Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0210243Z return mod(**inputs) 2025-09-07T07:41:15.0210612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0210701Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0211068Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0211160Z layer_outputs = layer_module( 2025-09-07T07:41:15.0211502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0211601Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0211965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0212071Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0212435Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0212539Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0212902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0213040Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0213049Z 2025-09-07T07:41:15.0213185Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0213469Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0213634Z return mod(**inputs) 2025-09-07T07:41:15.0214008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0214099Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0214469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0214553Z layer_outputs = layer_module( 2025-09-07T07:41:15.0214895Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0214991Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0215356Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0215518Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0215885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0215992Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0216355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0216498Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0216503Z 2025-09-07T07:41:15.0216635Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0216920Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0216997Z return mod(**inputs) 2025-09-07T07:41:15.0217361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0217456Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0217823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0217914Z layer_outputs = layer_module( 2025-09-07T07:41:15.0218255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0218351Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0218715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0218816Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0219181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0219284Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0219650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0219792Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0219797Z 2025-09-07T07:41:15.0219934Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0220223Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0220299Z return mod(**inputs) 2025-09-07T07:41:15.0220664Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0220756Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0221124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0221213Z layer_outputs = layer_module( 2025-09-07T07:41:15.0221556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0221658Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0222021Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0222214Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0222585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0222688Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0223055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:15.0223194Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:15.0223199Z 2025-09-07T07:41:15.0223332Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0223618Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0223752Z return mod(**inputs) 2025-09-07T07:41:15.0224123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0224217Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0224587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0224671Z layer_outputs = layer_module( 2025-09-07T07:41:15.0225010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0225108Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0225471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0225573Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0225938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0226043Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0226413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:15.0226504Z attn_output = self.o(attn_output) 2025-09-07T07:41:15.0226508Z 2025-09-07T07:41:15.0226647Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0226932Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0227007Z return mod(**inputs) 2025-09-07T07:41:15.0227377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0227464Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0227833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0227922Z layer_outputs = layer_module( 2025-09-07T07:41:15.0228267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0228364Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0228731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0228834Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0229195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T07:41:15.0229373Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T07:41:15.0229378Z 2025-09-07T07:41:15.0229474Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0229607Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0229894Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0229976Z return mod(**inputs) 2025-09-07T07:41:15.0230345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0230521Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0230893Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0230981Z layer_outputs = layer_module( 2025-09-07T07:41:15.0231322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0231421Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0231784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0231898Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0232262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:41:15.0232460Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0232831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0232924Z return self.weight * hidden_states 2025-09-07T07:41:15.0232929Z 2025-09-07T07:41:15.0233065Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0233350Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0233426Z return mod(**inputs) 2025-09-07T07:41:15.0233796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0233885Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0234253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0234343Z layer_outputs = layer_module( 2025-09-07T07:41:15.0234688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0234788Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0235149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0235263Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0235625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0235780Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0236142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:41:15.0236237Z hidden_states = self.wi(hidden_states) 2025-09-07T07:41:15.0236245Z 2025-09-07T07:41:15.0236382Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0236666Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0236745Z return mod(**inputs) 2025-09-07T07:41:15.0237115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0237204Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0237573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0237660Z layer_outputs = layer_module( 2025-09-07T07:41:15.0238004Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0238099Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0238465Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0238579Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0239057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0239218Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0239582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:41:15.0239683Z hidden_states = self.act(hidden_states) 2025-09-07T07:41:15.0239688Z 2025-09-07T07:41:15.0239820Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0240105Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0240183Z return mod(**inputs) 2025-09-07T07:41:15.0240550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0240709Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0241075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0241164Z layer_outputs = layer_module( 2025-09-07T07:41:15.0241504Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0241600Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0241964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0242074Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0242440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0242589Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0242955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:41:15.0243053Z hidden_states = self.wo(hidden_states) 2025-09-07T07:41:15.0243057Z 2025-09-07T07:41:15.0243156Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0243296Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0243580Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0243655Z return mod(**inputs) 2025-09-07T07:41:15.0244025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0244114Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0244483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0244567Z layer_outputs = layer_module( 2025-09-07T07:41:15.0244914Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0245009Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0245377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0245482Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0245845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:41:15.0245982Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0246346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0246437Z return self.weight * hidden_states 2025-09-07T07:41:15.0246441Z 2025-09-07T07:41:15.0246579Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0246863Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0246949Z return mod(**inputs) 2025-09-07T07:41:15.0247315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0247496Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0247870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0247956Z layer_outputs = layer_module( 2025-09-07T07:41:15.0248302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0248396Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0248760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0248861Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0249339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0249443Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0249807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:15.0249902Z query_states = self.q(hidden_states) 2025-09-07T07:41:15.0249907Z 2025-09-07T07:41:15.0250042Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0250326Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0250404Z return mod(**inputs) 2025-09-07T07:41:15.0250769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0250858Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0251224Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0251316Z layer_outputs = layer_module( 2025-09-07T07:41:15.0251657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0251755Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0252123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0252223Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0252587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0252689Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0253051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:15.0253145Z key_states = self.k(current_states) 2025-09-07T07:41:15.0253154Z 2025-09-07T07:41:15.0253288Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0253577Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0253654Z return mod(**inputs) 2025-09-07T07:41:15.0254025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0254115Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0254480Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0254568Z layer_outputs = layer_module( 2025-09-07T07:41:15.0254909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0255009Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0255372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0255475Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0255839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0256020Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0256396Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0256565Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0256570Z 2025-09-07T07:41:15.0256704Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0256995Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0257069Z return mod(**inputs) 2025-09-07T07:41:15.0257440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0257587Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0257957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0258045Z layer_outputs = layer_module( 2025-09-07T07:41:15.0258387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0258487Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0258853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0258956Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0259320Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0259421Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0259787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0259960Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0259965Z 2025-09-07T07:41:15.0260104Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0260389Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0260465Z return mod(**inputs) 2025-09-07T07:41:15.0260835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0260922Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0261294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0261380Z layer_outputs = layer_module( 2025-09-07T07:41:15.0261721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0261820Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0262182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0262288Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0262650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0262755Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0263118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0263285Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0263292Z 2025-09-07T07:41:15.0263427Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0263714Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0263795Z return mod(**inputs) 2025-09-07T07:41:15.0264160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0264330Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0264703Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0264789Z layer_outputs = layer_module( 2025-09-07T07:41:15.0265133Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0265228Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0265667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0265772Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0266137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0266384Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0266753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0266967Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0266973Z 2025-09-07T07:41:15.0267107Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0267395Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0267471Z return mod(**inputs) 2025-09-07T07:41:15.0267843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0267936Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0268303Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0268396Z layer_outputs = layer_module( 2025-09-07T07:41:15.0268737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0268837Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0269208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0269309Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0269675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0269777Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0270141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0270352Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0270361Z 2025-09-07T07:41:15.0270496Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0270786Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0270864Z return mod(**inputs) 2025-09-07T07:41:15.0271232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0271321Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0271687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0271774Z layer_outputs = layer_module( 2025-09-07T07:41:15.0272117Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0272216Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0272579Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0272683Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0273180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0273292Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0273659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:15.0273753Z value_states = self.v(current_states) 2025-09-07T07:41:15.0273758Z 2025-09-07T07:41:15.0273895Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0274182Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0274258Z return mod(**inputs) 2025-09-07T07:41:15.0274629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0274785Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0275153Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0275241Z layer_outputs = layer_module( 2025-09-07T07:41:15.0275580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0275678Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0276039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0276142Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0276503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0276603Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0276968Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0277109Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0277114Z 2025-09-07T07:41:15.0277256Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0277541Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0277618Z return mod(**inputs) 2025-09-07T07:41:15.0277984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0278073Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0278443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0278528Z layer_outputs = layer_module( 2025-09-07T07:41:15.0278873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0278972Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0279334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0279439Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0279802Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0279904Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0280264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0280401Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0280409Z 2025-09-07T07:41:15.0280542Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0280829Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0280913Z return mod(**inputs) 2025-09-07T07:41:15.0281279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0281370Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0281826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0281917Z layer_outputs = layer_module( 2025-09-07T07:41:15.0282261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0282357Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0282723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0282825Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0283187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0283378Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0283742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0283886Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0283891Z 2025-09-07T07:41:15.0284028Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0284318Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0284394Z return mod(**inputs) 2025-09-07T07:41:15.0284762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0284852Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0285219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0285311Z layer_outputs = layer_module( 2025-09-07T07:41:15.0285652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0285746Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0286118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0286217Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0286584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0286684Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0287045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:15.0287186Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:15.0287191Z 2025-09-07T07:41:15.0287330Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0287621Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0287697Z return mod(**inputs) 2025-09-07T07:41:15.0288075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0288645Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0289197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0289749Z layer_outputs = layer_module( 2025-09-07T07:41:15.0290256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0290787Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0291338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0291904Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0292459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0293129Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0293698Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:15.0294251Z attn_output = self.o(attn_output) 2025-09-07T07:41:15.0294434Z 2025-09-07T07:41:15.0294533Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0294850Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0295366Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0295825Z return mod(**inputs) 2025-09-07T07:41:15.0296333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0296996Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0297567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0298117Z layer_outputs = layer_module( 2025-09-07T07:41:15.0298620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0299156Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0299708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0300277Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0300847Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:41:15.0301429Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0302017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0302570Z return self.weight * hidden_states 2025-09-07T07:41:15.0302755Z 2025-09-07T07:41:15.0302895Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0303411Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0303866Z return mod(**inputs) 2025-09-07T07:41:15.0304373Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0304924Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0305466Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0306013Z layer_outputs = layer_module( 2025-09-07T07:41:15.0306516Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0307053Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0307606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0308180Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0308749Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0309361Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0309974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:41:15.0310527Z hidden_states = self.wi(hidden_states) 2025-09-07T07:41:15.0310713Z 2025-09-07T07:41:15.0310851Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0311362Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0311825Z return mod(**inputs) 2025-09-07T07:41:15.0312332Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0312884Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0313532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0314092Z layer_outputs = layer_module( 2025-09-07T07:41:15.0314596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0315127Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0315680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0316242Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0316813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0317546Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0318166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:41:15.0318726Z hidden_states = self.act(hidden_states) 2025-09-07T07:41:15.0318917Z 2025-09-07T07:41:15.0319050Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0319566Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0320021Z return mod(**inputs) 2025-09-07T07:41:15.0320525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0321075Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0321612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0322166Z layer_outputs = layer_module( 2025-09-07T07:41:15.0322672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0323209Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0323760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0324326Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0324894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0325503Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0326111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:41:15.0326666Z hidden_states = self.wo(hidden_states) 2025-09-07T07:41:15.0326863Z 2025-09-07T07:41:15.0326960Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0327277Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0327792Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0328252Z return mod(**inputs) 2025-09-07T07:41:15.0328760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0329313Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0329851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0330400Z layer_outputs = layer_module( 2025-09-07T07:41:15.0330906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0331438Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0331995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0332556Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0333218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:41:15.0333822Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0334417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0334969Z return self.weight * hidden_states 2025-09-07T07:41:15.0335150Z 2025-09-07T07:41:15.0335285Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0335799Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0336253Z return mod(**inputs) 2025-09-07T07:41:15.0336755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0337401Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0337937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0338488Z layer_outputs = layer_module( 2025-09-07T07:41:15.0338992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0339522Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0340072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0340630Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0341183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0341745Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0342308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:15.0342861Z query_states = self.q(hidden_states) 2025-09-07T07:41:15.0343046Z 2025-09-07T07:41:15.0343185Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0343701Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0344158Z return mod(**inputs) 2025-09-07T07:41:15.0344665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0345212Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0345750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0346297Z layer_outputs = layer_module( 2025-09-07T07:41:15.0346800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0347339Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0347886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0348450Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0349012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0349570Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0350128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:15.0350678Z key_states = self.k(current_states) 2025-09-07T07:41:15.0350860Z 2025-09-07T07:41:15.0351000Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0351513Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0351972Z return mod(**inputs) 2025-09-07T07:41:15.0352476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0353024Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0353676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0354231Z layer_outputs = layer_module( 2025-09-07T07:41:15.0354736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0355269Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0355819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0356377Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0356925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0357570Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0358126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0358754Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0359018Z 2025-09-07T07:41:15.0359154Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0359665Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0360121Z return mod(**inputs) 2025-09-07T07:41:15.0360626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0361176Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0361709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0362258Z layer_outputs = layer_module( 2025-09-07T07:41:15.0362761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0363297Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0363848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0364401Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0364955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0365626Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0366192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0366815Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0367087Z 2025-09-07T07:41:15.0367221Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0367730Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0368186Z return mod(**inputs) 2025-09-07T07:41:15.0368693Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0369241Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0369781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0370330Z layer_outputs = layer_module( 2025-09-07T07:41:15.0370833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0371369Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0371916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0372477Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0373033Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0373766Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0374334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0374960Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0375228Z 2025-09-07T07:41:15.0375362Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0375877Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0376336Z return mod(**inputs) 2025-09-07T07:41:15.0376840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0377521Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0378060Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0378608Z layer_outputs = layer_module( 2025-09-07T07:41:15.0379119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0379649Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0380204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0380763Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0381320Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0381880Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0382436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0383107Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0383418Z 2025-09-07T07:41:15.0383560Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0384079Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0384533Z return mod(**inputs) 2025-09-07T07:41:15.0385038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0385594Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0386133Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0386680Z layer_outputs = layer_module( 2025-09-07T07:41:15.0387180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0387719Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0388272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0388835Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0389389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0389951Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0390511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0391178Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0391485Z 2025-09-07T07:41:15.0391623Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0392132Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0392593Z return mod(**inputs) 2025-09-07T07:41:15.0393096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0393783Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0394335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0394883Z layer_outputs = layer_module( 2025-09-07T07:41:15.0395387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0395922Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0396476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0397032Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0397586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0398236Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0398799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:15.0399352Z value_states = self.v(current_states) 2025-09-07T07:41:15.0399540Z 2025-09-07T07:41:15.0399675Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0400191Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0400652Z return mod(**inputs) 2025-09-07T07:41:15.0401161Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0401714Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0402252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0402809Z layer_outputs = layer_module( 2025-09-07T07:41:15.0403317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0403855Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0404410Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0404970Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0405525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0406088Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0406645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0407239Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0407483Z 2025-09-07T07:41:15.0407615Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0408128Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0408587Z return mod(**inputs) 2025-09-07T07:41:15.0409093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0409645Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0410184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0410729Z layer_outputs = layer_module( 2025-09-07T07:41:15.0411237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0411770Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0412327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0412893Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0413454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0414130Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0414699Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0415299Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0415537Z 2025-09-07T07:41:15.0415669Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0416183Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0416635Z return mod(**inputs) 2025-09-07T07:41:15.0417138Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0417778Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0418315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0418864Z layer_outputs = layer_module( 2025-09-07T07:41:15.0419370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0419905Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0420459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0421019Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0421575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0422133Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0422696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0423296Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0423531Z 2025-09-07T07:41:15.0423667Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0424180Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0424639Z return mod(**inputs) 2025-09-07T07:41:15.0425146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0425698Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0426240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0426783Z layer_outputs = layer_module( 2025-09-07T07:41:15.0427287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0427827Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0428383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0428943Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0429498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0430058Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0430615Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:15.0431214Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:15.0431451Z 2025-09-07T07:41:15.0431584Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0432099Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0432563Z return mod(**inputs) 2025-09-07T07:41:15.0433071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0433626Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0434275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0434838Z layer_outputs = layer_module( 2025-09-07T07:41:15.0435345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0435881Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0436434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0436989Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0437541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0438193Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0438753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:15.0439304Z attn_output = self.o(attn_output) 2025-09-07T07:41:15.0439484Z 2025-09-07T07:41:15.0439620Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0440136Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0440594Z return mod(**inputs) 2025-09-07T07:41:15.0441101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0441649Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0442191Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0442740Z layer_outputs = layer_module( 2025-09-07T07:41:15.0443247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0443777Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0444333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0444893Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0445451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T07:41:15.0446083Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T07:41:15.0446356Z 2025-09-07T07:41:15.0446454Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0446774Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0455628Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0456124Z return mod(**inputs) 2025-09-07T07:41:15.0456637Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0457196Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0457748Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0458296Z layer_outputs = layer_module( 2025-09-07T07:41:15.0458804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0459339Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0459892Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0460461Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0461031Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:41:15.0461617Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0462202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0462920Z return self.weight * hidden_states 2025-09-07T07:41:15.0463124Z 2025-09-07T07:41:15.0463261Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0463776Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0464233Z return mod(**inputs) 2025-09-07T07:41:15.0464741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0465382Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0465919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0466620Z layer_outputs = layer_module( 2025-09-07T07:41:15.0467126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0467660Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0468214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0468785Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0469354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0469966Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0470577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:41:15.0471128Z hidden_states = self.wi(hidden_states) 2025-09-07T07:41:15.0471321Z 2025-09-07T07:41:15.0471460Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0471973Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0472431Z return mod(**inputs) 2025-09-07T07:41:15.0472937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0473488Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0474027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0474579Z layer_outputs = layer_module( 2025-09-07T07:41:15.0475083Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0475611Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0476163Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0476738Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0477305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0477917Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0478527Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:41:15.0479083Z hidden_states = self.act(hidden_states) 2025-09-07T07:41:15.0479275Z 2025-09-07T07:41:15.0479409Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0479923Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0480382Z return mod(**inputs) 2025-09-07T07:41:15.0480883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0481439Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0481975Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0482520Z layer_outputs = layer_module( 2025-09-07T07:41:15.0483213Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0483756Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0484308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0484875Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0485445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0486049Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0486660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:41:15.0487305Z hidden_states = self.wo(hidden_states) 2025-09-07T07:41:15.0487493Z 2025-09-07T07:41:15.0487592Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0487907Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0488417Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0488875Z return mod(**inputs) 2025-09-07T07:41:15.0489381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:41:15.0489928Z encoder_outputs = self.encoder( 2025-09-07T07:41:15.0490463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1128, in forward 2025-09-07T07:41:15.0491054Z hidden_states = self.final_layer_norm(hidden_states) 2025-09-07T07:41:15.0491641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0492193Z return self.weight * hidden_states 2025-09-07T07:41:15.0492374Z 2025-09-07T07:41:15.0492511Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0493023Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0493481Z return mod(**inputs) 2025-09-07T07:41:15.0493986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0494536Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0495070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0495616Z layer_outputs = layer_module( 2025-09-07T07:41:15.0496120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0496657Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0497208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0497771Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0498334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0498899Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0499462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:15.0500007Z key_states = self.k(current_states) 2025-09-07T07:41:15.0500192Z 2025-09-07T07:41:15.0500325Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0500840Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0501300Z return mod(**inputs) 2025-09-07T07:41:15.0501809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0502354Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0503007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0503574Z layer_outputs = layer_module( 2025-09-07T07:41:15.0504083Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0504612Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0505160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0505721Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0506278Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0506934Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0507489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0508121Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0508391Z 2025-09-07T07:41:15.0508525Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0509035Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0509490Z return mod(**inputs) 2025-09-07T07:41:15.0509988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0510535Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0511069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0511618Z layer_outputs = layer_module( 2025-09-07T07:41:15.0512124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0512655Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0513216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0513776Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0514332Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0514890Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0515451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0516073Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0516337Z 2025-09-07T07:41:15.0516475Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0516993Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0517449Z return mod(**inputs) 2025-09-07T07:41:15.0517953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0518507Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0519049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0519594Z layer_outputs = layer_module( 2025-09-07T07:41:15.0520096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0520628Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0521180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0521742Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0522294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0522983Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0523561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0524193Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0524459Z 2025-09-07T07:41:15.0524598Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0525108Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0525563Z return mod(**inputs) 2025-09-07T07:41:15.0526074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0526751Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0527292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0527837Z layer_outputs = layer_module( 2025-09-07T07:41:15.0528343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0528873Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0529421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0529977Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0530532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0531097Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0531655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0532326Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0532633Z 2025-09-07T07:41:15.0532768Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0533288Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0533742Z return mod(**inputs) 2025-09-07T07:41:15.0534249Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0534799Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0535331Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0535880Z layer_outputs = layer_module( 2025-09-07T07:41:15.0536384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0536919Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0537467Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0538031Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0538593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0539162Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0539718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:15.0540265Z value_states = self.v(current_states) 2025-09-07T07:41:15.0540451Z 2025-09-07T07:41:15.0540584Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0541095Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0541554Z return mod(**inputs) 2025-09-07T07:41:15.0542064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0542614Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0543282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0543841Z layer_outputs = layer_module( 2025-09-07T07:41:15.0544344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0544874Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0545424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0545983Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0546538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0547222Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0547780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0548384Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0548622Z 2025-09-07T07:41:15.0548756Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0549271Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0549724Z return mod(**inputs) 2025-09-07T07:41:15.0550227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0550779Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0551316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0551873Z layer_outputs = layer_module( 2025-09-07T07:41:15.0552378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0552916Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0553470Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0554027Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0554585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0555150Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0555715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0556312Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0556544Z 2025-09-07T07:41:15.0556679Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0557194Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0557650Z return mod(**inputs) 2025-09-07T07:41:15.0558159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0558709Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0559246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0559792Z layer_outputs = layer_module( 2025-09-07T07:41:15.0560294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0560826Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0561376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0561948Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0562504Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0563068Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0563756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0563904Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0563910Z 2025-09-07T07:41:15.0564043Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0564333Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0564408Z return mod(**inputs) 2025-09-07T07:41:15.0564779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0564868Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0565386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0565477Z layer_outputs = layer_module( 2025-09-07T07:41:15.0565822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0565920Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0566285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0566386Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0566750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0566857Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0567222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:15.0567365Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:15.0567369Z 2025-09-07T07:41:15.0567505Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0567796Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0567870Z return mod(**inputs) 2025-09-07T07:41:15.0568240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0568328Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0568697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0568781Z layer_outputs = layer_module( 2025-09-07T07:41:15.0569122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0569220Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0569589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0569694Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0570060Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0570167Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0570529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:15.0570618Z attn_output = self.o(attn_output) 2025-09-07T07:41:15.0570623Z 2025-09-07T07:41:15.0570722Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0570856Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0571143Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0571224Z return mod(**inputs) 2025-09-07T07:41:15.0571592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0571686Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0572216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0572314Z layer_outputs = layer_module( 2025-09-07T07:41:15.0572652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0572748Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0573113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0573225Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0573591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:41:15.0573829Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0574197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0574295Z return self.weight * hidden_states 2025-09-07T07:41:15.0574300Z 2025-09-07T07:41:15.0574435Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0574720Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0574796Z return mod(**inputs) 2025-09-07T07:41:15.0575166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0575255Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0575619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0575714Z layer_outputs = layer_module( 2025-09-07T07:41:15.0576056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0576155Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0576523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0576634Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0577000Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0577152Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0577516Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:41:15.0577610Z hidden_states = self.wi(hidden_states) 2025-09-07T07:41:15.0577615Z 2025-09-07T07:41:15.0577749Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0578037Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0578111Z return mod(**inputs) 2025-09-07T07:41:15.0578486Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0578575Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0578942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0579030Z layer_outputs = layer_module( 2025-09-07T07:41:15.0579369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0579467Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0579830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0579950Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0580316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0580565Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0580937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:41:15.0581035Z hidden_states = self.act(hidden_states) 2025-09-07T07:41:15.0581040Z 2025-09-07T07:41:15.0581175Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0581461Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0581541Z return mod(**inputs) 2025-09-07T07:41:15.0581907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0581999Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0582451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0582536Z layer_outputs = layer_module( 2025-09-07T07:41:15.0582884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0582981Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0583344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0583457Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0583821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0583975Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0584338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:41:15.0584441Z hidden_states = self.wo(hidden_states) 2025-09-07T07:41:15.0584446Z 2025-09-07T07:41:15.0584578Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0584866Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0584945Z return mod(**inputs) 2025-09-07T07:41:15.0585317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0585411Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0585780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0585865Z layer_outputs = layer_module( 2025-09-07T07:41:15.0586209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0586309Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0586675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0586774Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0587142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:41:15.0587276Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0587640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0587737Z return self.weight * hidden_states 2025-09-07T07:41:15.0587742Z 2025-09-07T07:41:15.0587877Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0588165Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0588241Z return mod(**inputs) 2025-09-07T07:41:15.0588612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0588704Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0589156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0589251Z layer_outputs = layer_module( 2025-09-07T07:41:15.0589593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0589688Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0590055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0590154Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0590521Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0590704Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0591069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:15.0591162Z query_states = self.q(hidden_states) 2025-09-07T07:41:15.0591170Z 2025-09-07T07:41:15.0591306Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0591597Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0591676Z return mod(**inputs) 2025-09-07T07:41:15.0592047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0592138Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0592505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0592594Z layer_outputs = layer_module( 2025-09-07T07:41:15.0592937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0593035Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0593403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0593506Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0593870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0593974Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0594342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:15.0594434Z key_states = self.k(current_states) 2025-09-07T07:41:15.0594439Z 2025-09-07T07:41:15.0594574Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0594865Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0594942Z return mod(**inputs) 2025-09-07T07:41:15.0595317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0595407Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0595775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0595862Z layer_outputs = layer_module( 2025-09-07T07:41:15.0596202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0596302Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0596666Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0596772Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0597140Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0597246Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0597706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0597887Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0597892Z 2025-09-07T07:41:15.0598032Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0598318Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0598397Z return mod(**inputs) 2025-09-07T07:41:15.0598764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0598853Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0599222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0599386Z layer_outputs = layer_module( 2025-09-07T07:41:15.0599734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0599829Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0600199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0600300Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0600664Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0600770Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0601132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0601305Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0601314Z 2025-09-07T07:41:15.0601448Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0601739Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0601819Z return mod(**inputs) 2025-09-07T07:41:15.0602188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0602279Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0602647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0602733Z layer_outputs = layer_module( 2025-09-07T07:41:15.0603078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0603175Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0603547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0603648Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0604016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0604120Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0604483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0604656Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0604661Z 2025-09-07T07:41:15.0604795Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0605084Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0605159Z return mod(**inputs) 2025-09-07T07:41:15.0605524Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0605621Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0606075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0606173Z layer_outputs = layer_module( 2025-09-07T07:41:15.0606514Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0606613Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0606978Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0607078Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0607446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0607549Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0607996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0608207Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0608216Z 2025-09-07T07:41:15.0608350Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0608636Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0608711Z return mod(**inputs) 2025-09-07T07:41:15.0609084Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0609174Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0609541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0609630Z layer_outputs = layer_module( 2025-09-07T07:41:15.0609975Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0610073Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0610438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0610541Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0610903Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0611005Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0611374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:15.0611466Z value_states = self.v(current_states) 2025-09-07T07:41:15.0611471Z 2025-09-07T07:41:15.0611609Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0611895Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0611975Z return mod(**inputs) 2025-09-07T07:41:15.0612346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0612442Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0612812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0612899Z layer_outputs = layer_module( 2025-09-07T07:41:15.0613242Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0613336Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0613699Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0613803Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0614171Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0614278Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0614732Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0614880Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0614885Z 2025-09-07T07:41:15.0615026Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0615311Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0615390Z return mod(**inputs) 2025-09-07T07:41:15.0615757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0615846Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0616217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0616369Z layer_outputs = layer_module( 2025-09-07T07:41:15.0616714Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0616816Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0617183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0617283Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0617644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0617750Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0618114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0618256Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0618265Z 2025-09-07T07:41:15.0618398Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0618682Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0618765Z return mod(**inputs) 2025-09-07T07:41:15.0619131Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0619224Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0619591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0619678Z layer_outputs = layer_module( 2025-09-07T07:41:15.0620019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0620114Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0620482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0620587Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0620956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0621061Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0621422Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0621566Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0621571Z 2025-09-07T07:41:15.0621704Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0621990Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0622068Z return mod(**inputs) 2025-09-07T07:41:15.0622433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0622529Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0622895Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0623090Z layer_outputs = layer_module( 2025-09-07T07:41:15.0623442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0623543Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0623905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0624005Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0624383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0624486Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0624923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:15.0625063Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:15.0625069Z 2025-09-07T07:41:15.0625207Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0625494Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0625570Z return mod(**inputs) 2025-09-07T07:41:15.0625941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0626031Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0626401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0626488Z layer_outputs = layer_module( 2025-09-07T07:41:15.0626829Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0626935Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0627299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0627408Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0627771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0627872Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0628239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:15.0628330Z attn_output = self.o(attn_output) 2025-09-07T07:41:15.0628334Z 2025-09-07T07:41:15.0628434Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0628571Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0628860Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0628945Z return mod(**inputs) 2025-09-07T07:41:15.0629324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0629425Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0629803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0629895Z layer_outputs = layer_module( 2025-09-07T07:41:15.0630243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0630341Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0630718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0630822Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0631206Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:41:15.0631346Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0631810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0631917Z return self.weight * hidden_states 2025-09-07T07:41:15.0631922Z 2025-09-07T07:41:15.0632061Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0632360Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0632440Z return mod(**inputs) 2025-09-07T07:41:15.0632815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0632908Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0633291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0633453Z layer_outputs = layer_module( 2025-09-07T07:41:15.0633806Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0633910Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0634289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0634394Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0634779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0634889Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0635280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:15.0635377Z query_states = self.q(hidden_states) 2025-09-07T07:41:15.0635386Z 2025-09-07T07:41:15.0635522Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0635818Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0635901Z return mod(**inputs) 2025-09-07T07:41:15.0636285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0636377Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0636759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0636851Z layer_outputs = layer_module( 2025-09-07T07:41:15.0637206Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0637314Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0637694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0637805Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0638188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0638298Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0638676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:15.0638770Z key_states = self.k(current_states) 2025-09-07T07:41:15.0638774Z 2025-09-07T07:41:15.0638913Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0639197Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0639273Z return mod(**inputs) 2025-09-07T07:41:15.0639643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0639736Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0640108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0640279Z layer_outputs = layer_module( 2025-09-07T07:41:15.0640630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0640726Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0641091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0641200Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0641562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0641674Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0642037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0642270Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0642275Z 2025-09-07T07:41:15.0642421Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0642707Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0642786Z return mod(**inputs) 2025-09-07T07:41:15.0643152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0643241Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0643611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0643696Z layer_outputs = layer_module( 2025-09-07T07:41:15.0644040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0644141Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0644508Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0644614Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0644979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0645088Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0645451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0645621Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0645626Z 2025-09-07T07:41:15.0645761Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0646044Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0646127Z return mod(**inputs) 2025-09-07T07:41:15.0646495Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0646589Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0646960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0647048Z layer_outputs = layer_module( 2025-09-07T07:41:15.0647388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0647484Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0647852Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0647955Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0648318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0648433Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0648885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0649062Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0649067Z 2025-09-07T07:41:15.0649201Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0649490Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0649564Z return mod(**inputs) 2025-09-07T07:41:15.0649932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0650023Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0650386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0650525Z layer_outputs = layer_module( 2025-09-07T07:41:15.0650867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0650968Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0651330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0651432Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0651798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0651905Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0652268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0652476Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0652486Z 2025-09-07T07:41:15.0652621Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0652910Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0652990Z return mod(**inputs) 2025-09-07T07:41:15.0653360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0653450Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0653817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0653902Z layer_outputs = layer_module( 2025-09-07T07:41:15.0654245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0654347Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0654709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0654820Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0655187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0655294Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0655665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:15.0655759Z value_states = self.v(current_states) 2025-09-07T07:41:15.0655764Z 2025-09-07T07:41:15.0655904Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0656192Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0656274Z return mod(**inputs) 2025-09-07T07:41:15.0656647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0656743Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0657116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0657323Z layer_outputs = layer_module( 2025-09-07T07:41:15.0657681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0657778Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0658144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0658252Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0658614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0658723Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0659085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0659304Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0659313Z 2025-09-07T07:41:15.0659446Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0659735Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0659812Z return mod(**inputs) 2025-09-07T07:41:15.0660178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0660269Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0660636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0660721Z layer_outputs = layer_module( 2025-09-07T07:41:15.0661067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0661167Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0661533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0661639Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0661999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0662107Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0662469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0662609Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0662614Z 2025-09-07T07:41:15.0662747Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0663033Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0663114Z return mod(**inputs) 2025-09-07T07:41:15.0663484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0663576Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0663952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0664043Z layer_outputs = layer_module( 2025-09-07T07:41:15.0664388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0664484Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0664859Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0664963Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0665425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0665544Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0665921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0666256Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0666262Z 2025-09-07T07:41:15.0666416Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0666715Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0666790Z return mod(**inputs) 2025-09-07T07:41:15.0667166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0667255Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0667626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0667835Z layer_outputs = layer_module( 2025-09-07T07:41:15.0668180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0668278Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0668647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0668750Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0669121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0669226Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0669590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:15.0669731Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:15.0669736Z 2025-09-07T07:41:15.0669877Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0670162Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0670239Z return mod(**inputs) 2025-09-07T07:41:15.0670613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0670700Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0671070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0671155Z layer_outputs = layer_module( 2025-09-07T07:41:15.0671496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0671594Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0671958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0672067Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0672430Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0672540Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0672916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:15.0673006Z attn_output = self.o(attn_output) 2025-09-07T07:41:15.0673011Z 2025-09-07T07:41:15.0673114Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0673249Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0673546Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0673624Z return mod(**inputs) 2025-09-07T07:41:15.0673996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0674102Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0674477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0674665Z layer_outputs = layer_module( 2025-09-07T07:41:15.0675022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0675120Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0675494Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0675607Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0675977Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:41:15.0676103Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0676548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0676647Z return self.weight * hidden_states 2025-09-07T07:41:15.0676652Z 2025-09-07T07:41:15.0676791Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0677081Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0677157Z return mod(**inputs) 2025-09-07T07:41:15.0677528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0677618Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0677988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0678078Z layer_outputs = layer_module( 2025-09-07T07:41:15.0678424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0678527Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0678892Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0679005Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0679377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0679529Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0679896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:41:15.0679992Z hidden_states = self.wi(hidden_states) 2025-09-07T07:41:15.0679997Z 2025-09-07T07:41:15.0680136Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0680423Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0680503Z return mod(**inputs) 2025-09-07T07:41:15.0680876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0680965Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0681338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0681425Z layer_outputs = layer_module( 2025-09-07T07:41:15.0681768Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0681869Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0682232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0682343Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0682707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0682864Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0683348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:41:15.0683457Z hidden_states = self.act(hidden_states) 2025-09-07T07:41:15.0683462Z 2025-09-07T07:41:15.0683601Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0683886Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0683968Z return mod(**inputs) 2025-09-07T07:41:15.0684334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0684422Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0684795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0684979Z layer_outputs = layer_module( 2025-09-07T07:41:15.0685323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0685423Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0685788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0685902Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0686263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0686418Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0686780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:41:15.0686878Z hidden_states = self.wo(hidden_states) 2025-09-07T07:41:15.0686887Z 2025-09-07T07:41:15.0686982Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0687117Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0687404Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0687486Z return mod(**inputs) 2025-09-07T07:41:15.0687857Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0687946Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0688314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0688402Z layer_outputs = layer_module( 2025-09-07T07:41:15.0688745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0688846Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0689215Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0689316Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0689684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:41:15.0689819Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0690189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0690284Z return self.weight * hidden_states 2025-09-07T07:41:15.0690290Z 2025-09-07T07:41:15.0690426Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0690709Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0690784Z return mod(**inputs) 2025-09-07T07:41:15.0691154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0691247Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0691616Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0691833Z layer_outputs = layer_module( 2025-09-07T07:41:15.0692186Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0692286Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0692648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0692750Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0693112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0693214Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0693659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:15.0693751Z query_states = self.q(hidden_states) 2025-09-07T07:41:15.0693757Z 2025-09-07T07:41:15.0693898Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0694185Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0694264Z return mod(**inputs) 2025-09-07T07:41:15.0694631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0694721Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0695091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0695175Z layer_outputs = layer_module( 2025-09-07T07:41:15.0695523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0695624Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0695987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0696093Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0696455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0696561Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0696926Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:15.0697015Z key_states = self.k(current_states) 2025-09-07T07:41:15.0697023Z 2025-09-07T07:41:15.0697157Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0697442Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0697548Z return mod(**inputs) 2025-09-07T07:41:15.0697916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0698008Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0698377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0698463Z layer_outputs = layer_module( 2025-09-07T07:41:15.0698807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0698903Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0699269Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0699370Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0699729Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0699841Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0700202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0700465Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0700472Z 2025-09-07T07:41:15.0700614Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0700905Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0700983Z return mod(**inputs) 2025-09-07T07:41:15.0701351Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0701444Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0701808Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0701980Z layer_outputs = layer_module( 2025-09-07T07:41:15.0702323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0702422Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0702788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0702889Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0703257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0703362Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0703725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0703897Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0703906Z 2025-09-07T07:41:15.0704040Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0704329Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0704405Z return mod(**inputs) 2025-09-07T07:41:15.0704778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0704868Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0705236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0705325Z layer_outputs = layer_module( 2025-09-07T07:41:15.0705667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0705768Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0706129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0706232Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0706597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0706705Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0707071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0707244Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0707248Z 2025-09-07T07:41:15.0707384Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0707668Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0707743Z return mod(**inputs) 2025-09-07T07:41:15.0708114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0708208Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0708579Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0708665Z layer_outputs = layer_module( 2025-09-07T07:41:15.0709098Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0709205Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0709569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0709674Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0710038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0710141Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0710505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0710779Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0710784Z 2025-09-07T07:41:15.0710927Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0711212Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0711289Z return mod(**inputs) 2025-09-07T07:41:15.0711655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0711749Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0712122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0712207Z layer_outputs = layer_module( 2025-09-07T07:41:15.0712549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0712649Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0713012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0713117Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0713477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0713583Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0713943Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:15.0714040Z value_states = self.v(current_states) 2025-09-07T07:41:15.0714044Z 2025-09-07T07:41:15.0714179Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0714462Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0714548Z return mod(**inputs) 2025-09-07T07:41:15.0714920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0715013Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0715381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0715469Z layer_outputs = layer_module( 2025-09-07T07:41:15.0715817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0715914Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0716282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0716382Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0716744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0716862Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0717229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0717475Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0717480Z 2025-09-07T07:41:15.0717622Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0717913Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0717989Z return mod(**inputs) 2025-09-07T07:41:15.0718357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0718451Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0718816Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0718998Z layer_outputs = layer_module( 2025-09-07T07:41:15.0719342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0719442Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0719810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0719910Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0720276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0720379Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0720747Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0720884Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0720889Z 2025-09-07T07:41:15.0721027Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0721316Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0721393Z return mod(**inputs) 2025-09-07T07:41:15.0721766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0721857Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0722223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0722313Z layer_outputs = layer_module( 2025-09-07T07:41:15.0722654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0722755Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0723119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0723222Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0723588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0723693Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0724060Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0724197Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0724201Z 2025-09-07T07:41:15.0724337Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0724623Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0724699Z return mod(**inputs) 2025-09-07T07:41:15.0725071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0725165Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0725534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0725622Z layer_outputs = layer_module( 2025-09-07T07:41:15.0726061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0726165Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0726531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0726633Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0726997Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0727103Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0727464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:15.0727672Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:15.0727678Z 2025-09-07T07:41:15.0727819Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0728110Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0728188Z return mod(**inputs) 2025-09-07T07:41:15.0728555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0728644Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0729019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0729105Z layer_outputs = layer_module( 2025-09-07T07:41:15.0729450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0729554Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0729921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0730026Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0730391Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0730497Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0730858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:15.0730954Z attn_output = self.o(attn_output) 2025-09-07T07:41:15.0730959Z 2025-09-07T07:41:15.0731091Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0731377Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0731457Z return mod(**inputs) 2025-09-07T07:41:15.0731830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0731922Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0732289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0732376Z layer_outputs = layer_module( 2025-09-07T07:41:15.0732721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0732816Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0733181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0733281Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0733648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T07:41:15.0733829Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T07:41:15.0733834Z 2025-09-07T07:41:15.0733932Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0734071Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0734442Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0734531Z return mod(**inputs) 2025-09-07T07:41:15.0734899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0734988Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0735359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0735445Z layer_outputs = layer_module( 2025-09-07T07:41:15.0735792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0735951Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0736316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0736427Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0736791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:41:15.0736932Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0737293Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0737390Z return self.weight * hidden_states 2025-09-07T07:41:15.0737395Z 2025-09-07T07:41:15.0737528Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0737817Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0737902Z return mod(**inputs) 2025-09-07T07:41:15.0738274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0738365Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0738734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0738820Z layer_outputs = layer_module( 2025-09-07T07:41:15.0739166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0739264Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0739631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0739733Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0740093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0740210Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0740573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:15.0740677Z query_states = self.q(hidden_states) 2025-09-07T07:41:15.0740681Z 2025-09-07T07:41:15.0740818Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0741107Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0741183Z return mod(**inputs) 2025-09-07T07:41:15.0741551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0741644Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0742011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0742103Z layer_outputs = layer_module( 2025-09-07T07:41:15.0742449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0742545Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0743012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0743123Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0743490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0743599Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0743965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:15.0744056Z key_states = self.k(current_states) 2025-09-07T07:41:15.0744061Z 2025-09-07T07:41:15.0744195Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0744564Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0744641Z return mod(**inputs) 2025-09-07T07:41:15.0745016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0745105Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0745470Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0745561Z layer_outputs = layer_module( 2025-09-07T07:41:15.0745901Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0746003Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0746369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0746472Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0746841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0746948Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0747315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0747487Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0747492Z 2025-09-07T07:41:15.0747630Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0747918Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0747995Z return mod(**inputs) 2025-09-07T07:41:15.0748367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0748457Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0748831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0748917Z layer_outputs = layer_module( 2025-09-07T07:41:15.0749263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0749366Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0749728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0749834Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0750197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0750308Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0750671Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0750844Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0750849Z 2025-09-07T07:41:15.0750987Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0751407Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0751494Z return mod(**inputs) 2025-09-07T07:41:15.0751867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0751957Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0752331Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0752416Z layer_outputs = layer_module( 2025-09-07T07:41:15.0752760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0752857Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0753298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0753405Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0753774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0753884Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0754245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0754416Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0754421Z 2025-09-07T07:41:15.0754555Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0754842Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0754922Z return mod(**inputs) 2025-09-07T07:41:15.0755296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0755390Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0755757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0755842Z layer_outputs = layer_module( 2025-09-07T07:41:15.0756185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0756282Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0756648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0756750Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0757114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0757223Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0757586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0757806Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0757812Z 2025-09-07T07:41:15.0757946Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0758237Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0758315Z return mod(**inputs) 2025-09-07T07:41:15.0758689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0758785Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0759153Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0759242Z layer_outputs = layer_module( 2025-09-07T07:41:15.0759588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0759684Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0760140Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0760248Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0760615Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0760721Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0761087Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:15.0761184Z value_states = self.v(current_states) 2025-09-07T07:41:15.0761189Z 2025-09-07T07:41:15.0761322Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0761684Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0761761Z return mod(**inputs) 2025-09-07T07:41:15.0762137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0762227Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0762592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0762681Z layer_outputs = layer_module( 2025-09-07T07:41:15.0763021Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0763122Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0763485Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0763591Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0763957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0764063Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0764431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0764571Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0764576Z 2025-09-07T07:41:15.0764714Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0765001Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0765076Z return mod(**inputs) 2025-09-07T07:41:15.0765566Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0765655Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0766034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0766119Z layer_outputs = layer_module( 2025-09-07T07:41:15.0766465Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0766566Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0766929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0767037Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0767400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0767510Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0767872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0768014Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0768019Z 2025-09-07T07:41:15.0768155Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0768610Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0768702Z return mod(**inputs) 2025-09-07T07:41:15.0769071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0769159Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0769530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0769614Z layer_outputs = layer_module( 2025-09-07T07:41:15.0769956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0770052Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0770535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0770638Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0771004Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0771115Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0771476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0771618Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0771623Z 2025-09-07T07:41:15.0771757Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0772042Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0772121Z return mod(**inputs) 2025-09-07T07:41:15.0772489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0772586Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0772957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0773041Z layer_outputs = layer_module( 2025-09-07T07:41:15.0773385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0773480Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0773843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0773946Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0774310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0774416Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0774781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:15.0774922Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:15.0774930Z 2025-09-07T07:41:15.0775063Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0775352Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0775427Z return mod(**inputs) 2025-09-07T07:41:15.0775797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0775888Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0776255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0776345Z layer_outputs = layer_module( 2025-09-07T07:41:15.0776691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0776789Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0777240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0777350Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0777720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0777826Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0778193Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:15.0778284Z attn_output = self.o(attn_output) 2025-09-07T07:41:15.0778288Z 2025-09-07T07:41:15.0778385Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0778525Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0778877Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0778958Z return mod(**inputs) 2025-09-07T07:41:15.0779334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0779424Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0779799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0779884Z layer_outputs = layer_module( 2025-09-07T07:41:15.0780229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0780326Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0780695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0780811Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0781171Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:41:15.0781301Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0781672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0781773Z return self.weight * hidden_states 2025-09-07T07:41:15.0781778Z 2025-09-07T07:41:15.0781912Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0782198Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0782276Z return mod(**inputs) 2025-09-07T07:41:15.0782642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0782733Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0783105Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0783191Z layer_outputs = layer_module( 2025-09-07T07:41:15.0783540Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0783635Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0784001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0784113Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0784477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0784630Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0784994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:41:15.0785099Z hidden_states = self.wi(hidden_states) 2025-09-07T07:41:15.0785104Z 2025-09-07T07:41:15.0785238Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0785642Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0785728Z return mod(**inputs) 2025-09-07T07:41:15.0786099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0786195Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0786561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0786651Z layer_outputs = layer_module( 2025-09-07T07:41:15.0786991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0787091Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0787524Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0787634Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0788005Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0788158Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0788525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:41:15.0788627Z hidden_states = self.act(hidden_states) 2025-09-07T07:41:15.0788633Z 2025-09-07T07:41:15.0788767Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0789055Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0789132Z return mod(**inputs) 2025-09-07T07:41:15.0789506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0789594Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0789963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0790052Z layer_outputs = layer_module( 2025-09-07T07:41:15.0790393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0790492Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0790856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0790971Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0791335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0791490Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0791862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:41:15.0791960Z hidden_states = self.wo(hidden_states) 2025-09-07T07:41:15.0791969Z 2025-09-07T07:41:15.0792068Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0792205Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0792492Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0792571Z return mod(**inputs) 2025-09-07T07:41:15.0792939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0793031Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0793397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0793490Z layer_outputs = layer_module( 2025-09-07T07:41:15.0793832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0793927Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0794367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0794478Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0794846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:41:15.0794980Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0795342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0795439Z return self.weight * hidden_states 2025-09-07T07:41:15.0795443Z 2025-09-07T07:41:15.0795632Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0795921Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0796000Z return mod(**inputs) 2025-09-07T07:41:15.0796375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0796468Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0796835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0796925Z layer_outputs = layer_module( 2025-09-07T07:41:15.0797268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0797366Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0797727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0797830Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0798199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0798317Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0798685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:15.0798779Z query_states = self.q(hidden_states) 2025-09-07T07:41:15.0798784Z 2025-09-07T07:41:15.0798918Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0799208Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0799285Z return mod(**inputs) 2025-09-07T07:41:15.0806771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0806958Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0807378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0807469Z layer_outputs = layer_module( 2025-09-07T07:41:15.0807817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0807917Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0808286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0808390Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0808758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0808861Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0809227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:15.0809322Z key_states = self.k(current_states) 2025-09-07T07:41:15.0809328Z 2025-09-07T07:41:15.0809463Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0809931Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0810012Z return mod(**inputs) 2025-09-07T07:41:15.0810385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0810475Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0810839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0810928Z layer_outputs = layer_module( 2025-09-07T07:41:15.0811268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0811367Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0811819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0811919Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0812290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0812392Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0812759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0812927Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0812931Z 2025-09-07T07:41:15.0813069Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0813357Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0813433Z return mod(**inputs) 2025-09-07T07:41:15.0813809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0813897Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0814269Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0814354Z layer_outputs = layer_module( 2025-09-07T07:41:15.0814695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0814795Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0815160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0815263Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0815623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0815734Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0816095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0816263Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0816271Z 2025-09-07T07:41:15.0816411Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0816698Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0816778Z return mod(**inputs) 2025-09-07T07:41:15.0817147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0817235Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0817606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0817691Z layer_outputs = layer_module( 2025-09-07T07:41:15.0818041Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0818137Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0818588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0818700Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0819063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0819170Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0819530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0819701Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0819706Z 2025-09-07T07:41:15.0819841Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0820225Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0820304Z return mod(**inputs) 2025-09-07T07:41:15.0820676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0820769Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0821136Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0821221Z layer_outputs = layer_module( 2025-09-07T07:41:15.0821572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0821669Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0822048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0822148Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0822530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0822635Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0823009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0823227Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0823233Z 2025-09-07T07:41:15.0823370Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0823669Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0823745Z return mod(**inputs) 2025-09-07T07:41:15.0824123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0824215Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0824600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0824691Z layer_outputs = layer_module( 2025-09-07T07:41:15.0825051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0825154Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0825532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0825635Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0826013Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0826115Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0826492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:15.0826591Z value_states = self.v(current_states) 2025-09-07T07:41:15.0826596Z 2025-09-07T07:41:15.0826734Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0827190Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0827273Z return mod(**inputs) 2025-09-07T07:41:15.0827655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0827747Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0828125Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0828215Z layer_outputs = layer_module( 2025-09-07T07:41:15.0828568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0828669Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0829120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0829226Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0829604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0829708Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0830084Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0830226Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0830230Z 2025-09-07T07:41:15.0830372Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0830668Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0830746Z return mod(**inputs) 2025-09-07T07:41:15.0831132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0831227Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0831602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0831689Z layer_outputs = layer_module( 2025-09-07T07:41:15.0832030Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0832127Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0832488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0832591Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0832955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0833059Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0833424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0833563Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0833568Z 2025-09-07T07:41:15.0833710Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0833997Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0834076Z return mod(**inputs) 2025-09-07T07:41:15.0834442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0834532Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0834900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0834986Z layer_outputs = layer_module( 2025-09-07T07:41:15.0835329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0835431Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0835880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0835988Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0836350Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0836457Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0836816Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0836958Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0836963Z 2025-09-07T07:41:15.0837099Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0837384Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0837534Z return mod(**inputs) 2025-09-07T07:41:15.0837899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0837996Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0838364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0838455Z layer_outputs = layer_module( 2025-09-07T07:41:15.0838794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0838889Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0839256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0839358Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0839729Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0839831Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0840205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:15.0840348Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:15.0840353Z 2025-09-07T07:41:15.0840486Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0840776Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0840852Z return mod(**inputs) 2025-09-07T07:41:15.0841218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0841310Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0841675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0841769Z layer_outputs = layer_module( 2025-09-07T07:41:15.0842109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0842216Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0842578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0842676Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0843046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0843148Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0843515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:15.0843604Z attn_output = self.o(attn_output) 2025-09-07T07:41:15.0843613Z 2025-09-07T07:41:15.0843710Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0843848Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0844212Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0844297Z return mod(**inputs) 2025-09-07T07:41:15.0844665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0844753Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0845123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0845208Z layer_outputs = layer_module( 2025-09-07T07:41:15.0845550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0845645Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0846074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0846176Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0846541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:41:15.0846680Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0847043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0847139Z return self.weight * hidden_states 2025-09-07T07:41:15.0847144Z 2025-09-07T07:41:15.0847279Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0847565Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0847644Z return mod(**inputs) 2025-09-07T07:41:15.0848010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0848107Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0848476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0848564Z layer_outputs = layer_module( 2025-09-07T07:41:15.0848901Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0848996Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0849361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0849463Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0849828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0849934Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0850302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:15.0850398Z query_states = self.q(hidden_states) 2025-09-07T07:41:15.0850403Z 2025-09-07T07:41:15.0850541Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0850832Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0850908Z return mod(**inputs) 2025-09-07T07:41:15.0851277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0851370Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0851736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0851827Z layer_outputs = layer_module( 2025-09-07T07:41:15.0852168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0852272Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0852704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0852811Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0853177Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0853284Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0853648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:15.0853741Z key_states = self.k(current_states) 2025-09-07T07:41:15.0853746Z 2025-09-07T07:41:15.0853880Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0854171Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0854322Z return mod(**inputs) 2025-09-07T07:41:15.0854694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0854786Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0855151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0855241Z layer_outputs = layer_module( 2025-09-07T07:41:15.0855593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0855694Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0856073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0856184Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0856564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0856678Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0857069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0857238Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0857243Z 2025-09-07T07:41:15.0857382Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0857668Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0857743Z return mod(**inputs) 2025-09-07T07:41:15.0858116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0858209Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0858583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0858673Z layer_outputs = layer_module( 2025-09-07T07:41:15.0859018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0859120Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0859484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0859591Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0859956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0860065Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0860429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0860599Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0860607Z 2025-09-07T07:41:15.0860744Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0861030Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0861186Z return mod(**inputs) 2025-09-07T07:41:15.0861563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0861652Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0862024Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0862110Z layer_outputs = layer_module( 2025-09-07T07:41:15.0862458Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0862558Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0862926Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0863095Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0863462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0863573Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0863937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0864113Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0864117Z 2025-09-07T07:41:15.0864256Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0864560Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0864642Z return mod(**inputs) 2025-09-07T07:41:15.0865017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0865114Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0865562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0865655Z layer_outputs = layer_module( 2025-09-07T07:41:15.0865996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0866091Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0866458Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0866559Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0866926Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0867036Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0867405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0867620Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0867625Z 2025-09-07T07:41:15.0867762Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0868051Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0868125Z return mod(**inputs) 2025-09-07T07:41:15.0868495Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0868584Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0868949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0869039Z layer_outputs = layer_module( 2025-09-07T07:41:15.0869383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0869486Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0870028Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0870141Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0870510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0870616Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0870983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:15.0871077Z value_states = self.v(current_states) 2025-09-07T07:41:15.0871081Z 2025-09-07T07:41:15.0871214Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0871503Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0871671Z return mod(**inputs) 2025-09-07T07:41:15.0872042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0872135Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0872504Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0872588Z layer_outputs = layer_module( 2025-09-07T07:41:15.0872926Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0873027Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0873393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0873499Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0873859Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0873968Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0874335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0874473Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0874478Z 2025-09-07T07:41:15.0874617Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0874903Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0874981Z return mod(**inputs) 2025-09-07T07:41:15.0875345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0875432Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0875800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0875889Z layer_outputs = layer_module( 2025-09-07T07:41:15.0876230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0876330Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0876690Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0876795Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0877154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0877261Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0877621Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0877756Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0877768Z 2025-09-07T07:41:15.0877902Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0878185Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0878264Z return mod(**inputs) 2025-09-07T07:41:15.0878727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0878827Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0879196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0879282Z layer_outputs = layer_module( 2025-09-07T07:41:15.0879628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0879724Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0880090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0880249Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0880611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0880725Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0881087Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0881224Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0881229Z 2025-09-07T07:41:15.0881362Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0881650Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0881728Z return mod(**inputs) 2025-09-07T07:41:15.0882093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0882192Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0882557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0882650Z layer_outputs = layer_module( 2025-09-07T07:41:15.0882990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0883085Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0883454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0883553Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0883917Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0884024Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0884386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:15.0884534Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:15.0884538Z 2025-09-07T07:41:15.0884678Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0884968Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0885044Z return mod(**inputs) 2025-09-07T07:41:15.0885414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0885504Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0885869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0885958Z layer_outputs = layer_module( 2025-09-07T07:41:15.0886300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0886403Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0886768Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0886981Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0887368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0887480Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0887842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:15.0887936Z attn_output = self.o(attn_output) 2025-09-07T07:41:15.0887942Z 2025-09-07T07:41:15.0888075Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0888362Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0888520Z return mod(**inputs) 2025-09-07T07:41:15.0888888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0888983Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0889352Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0889441Z layer_outputs = layer_module( 2025-09-07T07:41:15.0889781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0889877Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0890244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0890346Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0890712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 647, in forward 2025-09-07T07:41:15.0890889Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-09-07T07:41:15.0890894Z 2025-09-07T07:41:15.0890994Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0891134Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0891420Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0891499Z return mod(**inputs) 2025-09-07T07:41:15.0891865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0891956Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0892321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0892407Z layer_outputs = layer_module( 2025-09-07T07:41:15.0892750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0892851Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0893216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0893332Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0893694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:41:15.0893822Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0894184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0894279Z return self.weight * hidden_states 2025-09-07T07:41:15.0894284Z 2025-09-07T07:41:15.0894420Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0894710Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0894790Z return mod(**inputs) 2025-09-07T07:41:15.0895159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0895345Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0895725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0895816Z layer_outputs = layer_module( 2025-09-07T07:41:15.0896158Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0896254Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0896620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0896733Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0897098Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0897330Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0897698Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:41:15.0897799Z hidden_states = self.wi(hidden_states) 2025-09-07T07:41:15.0897803Z 2025-09-07T07:41:15.0897939Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0898228Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0898303Z return mod(**inputs) 2025-09-07T07:41:15.0898676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0898764Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0899130Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0899225Z layer_outputs = layer_module( 2025-09-07T07:41:15.0899564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0899667Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0900030Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0900140Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0900506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0900658Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0901023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:41:15.0901122Z hidden_states = self.act(hidden_states) 2025-09-07T07:41:15.0901130Z 2025-09-07T07:41:15.0901268Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0901552Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0901628Z return mod(**inputs) 2025-09-07T07:41:15.0902002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0902093Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0902459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0902545Z layer_outputs = layer_module( 2025-09-07T07:41:15.0902883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0902982Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0903347Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0903465Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0903826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0904065Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0904440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:41:15.0904536Z hidden_states = self.wo(hidden_states) 2025-09-07T07:41:15.0904541Z 2025-09-07T07:41:15.0904641Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0904776Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0905066Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0905144Z return mod(**inputs) 2025-09-07T07:41:15.0905508Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0905702Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0906073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0906162Z layer_outputs = layer_module( 2025-09-07T07:41:15.0906502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0906598Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0906966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0907066Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0907431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:41:15.0907568Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0907937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0908031Z return self.weight * hidden_states 2025-09-07T07:41:15.0908035Z 2025-09-07T07:41:15.0908174Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0908463Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0908541Z return mod(**inputs) 2025-09-07T07:41:15.0908912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0909004Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0909374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0909464Z layer_outputs = layer_module( 2025-09-07T07:41:15.0909807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0909914Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0910280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0910380Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0910746Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0910849Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0911215Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:15.0911308Z query_states = self.q(hidden_states) 2025-09-07T07:41:15.0911314Z 2025-09-07T07:41:15.0911451Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0911735Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0911817Z return mod(**inputs) 2025-09-07T07:41:15.0912189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0912369Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0912749Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0912835Z layer_outputs = layer_module( 2025-09-07T07:41:15.0913177Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0913278Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0913638Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0913742Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0914103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0914279Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0914650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:15.0914741Z key_states = self.k(current_states) 2025-09-07T07:41:15.0914746Z 2025-09-07T07:41:15.0914884Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0915169Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0915248Z return mod(**inputs) 2025-09-07T07:41:15.0915616Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0915704Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0916074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0916164Z layer_outputs = layer_module( 2025-09-07T07:41:15.0916508Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0916610Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0916972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0917075Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0917436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0917540Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0917905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0918078Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0918087Z 2025-09-07T07:41:15.0918224Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0918509Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0918588Z return mod(**inputs) 2025-09-07T07:41:15.0918958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0919050Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0919416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0919503Z layer_outputs = layer_module( 2025-09-07T07:41:15.0919848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0919944Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0920310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0920415Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0920776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0920985Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0921354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0921526Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0921531Z 2025-09-07T07:41:15.0921668Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0921956Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0922031Z return mod(**inputs) 2025-09-07T07:41:15.0922397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0922567Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0922932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0923025Z layer_outputs = layer_module( 2025-09-07T07:41:15.0923367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0923463Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0923830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0923930Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0924295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0924398Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0924761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0924932Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0924937Z 2025-09-07T07:41:15.0925075Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0925366Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0925444Z return mod(**inputs) 2025-09-07T07:41:15.0925813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0925902Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0926267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0926357Z layer_outputs = layer_module( 2025-09-07T07:41:15.0926699Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0926803Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0927165Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0927269Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0927635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0927735Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0928105Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0928315Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0928320Z 2025-09-07T07:41:15.0928460Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0928745Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0928826Z return mod(**inputs) 2025-09-07T07:41:15.0929196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0929379Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0929759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0929845Z layer_outputs = layer_module( 2025-09-07T07:41:15.0930189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0930291Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0930656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0930760Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0931122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0931303Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0931670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:15.0931765Z value_states = self.v(current_states) 2025-09-07T07:41:15.0931769Z 2025-09-07T07:41:15.0931907Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0932192Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0932270Z return mod(**inputs) 2025-09-07T07:41:15.0932634Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0932723Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0933093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0933183Z layer_outputs = layer_module( 2025-09-07T07:41:15.0933523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0933624Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0933988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0934090Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0934449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0934557Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0934918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0935061Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0935070Z 2025-09-07T07:41:15.0935205Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0935492Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0935573Z return mod(**inputs) 2025-09-07T07:41:15.0935942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0936034Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0936401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0936484Z layer_outputs = layer_module( 2025-09-07T07:41:15.0936828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0936923Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0937288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0937393Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0937756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0937948Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0938317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0938459Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0938464Z 2025-09-07T07:41:15.0938598Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0938888Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0938965Z return mod(**inputs) 2025-09-07T07:41:15.0939332Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0939493Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0939858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0939947Z layer_outputs = layer_module( 2025-09-07T07:41:15.0940294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0940391Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0940758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0940858Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0941224Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0941326Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0941689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0941832Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0941837Z 2025-09-07T07:41:15.0941970Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0942262Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0942338Z return mod(**inputs) 2025-09-07T07:41:15.0942706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0942796Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0943162Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0943250Z layer_outputs = layer_module( 2025-09-07T07:41:15.0943592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0943694Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0944058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0944162Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0944527Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0944628Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0944995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:15.0945134Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:15.0945139Z 2025-09-07T07:41:15.0945278Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0945564Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0945646Z return mod(**inputs) 2025-09-07T07:41:15.0946021Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0946111Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0946570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0946668Z layer_outputs = layer_module( 2025-09-07T07:41:15.0947011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0947111Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0947474Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.0947578Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.0947939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.0948139Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.0948500Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:15.0948596Z attn_output = self.o(attn_output) 2025-09-07T07:41:15.0948600Z 2025-09-07T07:41:15.0948704Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0948840Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0949126Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0949202Z return mod(**inputs) 2025-09-07T07:41:15.0949568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0949661Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0950026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0950118Z layer_outputs = layer_module( 2025-09-07T07:41:15.0950459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0950559Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0950923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0951025Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0951390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:41:15.0951526Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0951890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0951984Z return self.weight * hidden_states 2025-09-07T07:41:15.0951989Z 2025-09-07T07:41:15.0952127Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0952420Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0952496Z return mod(**inputs) 2025-09-07T07:41:15.0952870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0952960Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0953324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0953414Z layer_outputs = layer_module( 2025-09-07T07:41:15.0953756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0953856Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0954218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0954329Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0954690Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0954888Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0955261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:15.0955354Z query_states = self.q(hidden_states) 2025-09-07T07:41:15.0955359Z 2025-09-07T07:41:15.0955496Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0955783Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0955858Z return mod(**inputs) 2025-09-07T07:41:15.0956229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0956416Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0956786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0956872Z layer_outputs = layer_module( 2025-09-07T07:41:15.0957213Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0957315Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0957677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0957783Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0958147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0958257Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0958619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:15.0958715Z key_states = self.k(current_states) 2025-09-07T07:41:15.0958720Z 2025-09-07T07:41:15.0958860Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0959148Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0959229Z return mod(**inputs) 2025-09-07T07:41:15.0959594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0959683Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0960054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0960139Z layer_outputs = layer_module( 2025-09-07T07:41:15.0960484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0960579Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0960950Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0961055Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0961420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0961531Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0961890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0962063Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0962068Z 2025-09-07T07:41:15.0962203Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0962489Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0962568Z return mod(**inputs) 2025-09-07T07:41:15.0962939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0963030Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0963471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0963562Z layer_outputs = layer_module( 2025-09-07T07:41:15.0963908Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0964003Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0964370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0964473Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0964838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0964997Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0965442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0965620Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0965625Z 2025-09-07T07:41:15.0965760Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0966048Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0966123Z return mod(**inputs) 2025-09-07T07:41:15.0966487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0966580Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0966944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0967032Z layer_outputs = layer_module( 2025-09-07T07:41:15.0967383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0967479Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0967845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0967949Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0968321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0968430Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0968796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.0968968Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.0968974Z 2025-09-07T07:41:15.0969109Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0969402Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0969479Z return mod(**inputs) 2025-09-07T07:41:15.0969855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0969945Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0970314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0970404Z layer_outputs = layer_module( 2025-09-07T07:41:15.0970744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0970845Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0971208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0971321Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0971682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0971789Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0972328Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.0972556Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.0972563Z 2025-09-07T07:41:15.0972704Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0972992Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0973069Z return mod(**inputs) 2025-09-07T07:41:15.0973439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0973648Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0974019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0974106Z layer_outputs = layer_module( 2025-09-07T07:41:15.0974455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0974553Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0974917Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0975023Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0975384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0975494Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0975858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:15.0975956Z value_states = self.v(current_states) 2025-09-07T07:41:15.0975960Z 2025-09-07T07:41:15.0976097Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0976386Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0976466Z return mod(**inputs) 2025-09-07T07:41:15.0976834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0976925Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0977294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0977380Z layer_outputs = layer_module( 2025-09-07T07:41:15.0977726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0977823Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0978195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0978296Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0978663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0978775Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0979136Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0979277Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0979282Z 2025-09-07T07:41:15.0979416Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0979701Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0979782Z return mod(**inputs) 2025-09-07T07:41:15.0980154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0980247Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0980706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0980804Z layer_outputs = layer_module( 2025-09-07T07:41:15.0981149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0981244Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0981610Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0981713Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0982074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0982289Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0982655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0982801Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0982807Z 2025-09-07T07:41:15.0982941Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0983230Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0983306Z return mod(**inputs) 2025-09-07T07:41:15.0983674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0983766Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0984132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0984222Z layer_outputs = layer_module( 2025-09-07T07:41:15.0984567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0984664Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0985029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0985132Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0985494Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0985600Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0985965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.0986106Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.0986111Z 2025-09-07T07:41:15.0986245Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0986540Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0986614Z return mod(**inputs) 2025-09-07T07:41:15.0986989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0987080Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0987451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0987537Z layer_outputs = layer_module( 2025-09-07T07:41:15.0987879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0987979Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0988343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0988447Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0988818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0988923Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0989380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:15.0989526Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:15.0989531Z 2025-09-07T07:41:15.0989669Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0989957Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0990034Z return mod(**inputs) 2025-09-07T07:41:15.0990403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0990491Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0990929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0991015Z layer_outputs = layer_module( 2025-09-07T07:41:15.0991361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0991457Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0991819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.0991924Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.0992291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.0992400Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.0992759Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:15.0992852Z attn_output = self.o(attn_output) 2025-09-07T07:41:15.0992857Z 2025-09-07T07:41:15.0992957Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.0993091Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0993383Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0993461Z return mod(**inputs) 2025-09-07T07:41:15.0993828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0993920Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0994285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0994374Z layer_outputs = layer_module( 2025-09-07T07:41:15.0994715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0994817Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0995179Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0995290Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0995660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:41:15.0995788Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.0996155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.0996249Z return self.weight * hidden_states 2025-09-07T07:41:15.0996254Z 2025-09-07T07:41:15.0996387Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.0996674Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.0996750Z return mod(**inputs) 2025-09-07T07:41:15.0997125Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.0997213Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.0997667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.0997764Z layer_outputs = layer_module( 2025-09-07T07:41:15.0998106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.0998208Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.0998569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.0998681Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.0999046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.0999266Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.0999633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:41:15.0999734Z hidden_states = self.wi(hidden_states) 2025-09-07T07:41:15.0999738Z 2025-09-07T07:41:15.0999876Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1000162Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1000238Z return mod(**inputs) 2025-09-07T07:41:15.1000609Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1000698Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1001067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1001154Z layer_outputs = layer_module( 2025-09-07T07:41:15.1001502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1001599Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1001965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.1002082Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.1002444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.1002600Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.1002962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:41:15.1003060Z hidden_states = self.act(hidden_states) 2025-09-07T07:41:15.1003064Z 2025-09-07T07:41:15.1003201Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1003490Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1003569Z return mod(**inputs) 2025-09-07T07:41:15.1003938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1004028Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1004394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1004479Z layer_outputs = layer_module( 2025-09-07T07:41:15.1004820Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1004915Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1005279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.1005394Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.1005756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.1006005Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.1006373Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:41:15.1006471Z hidden_states = self.wo(hidden_states) 2025-09-07T07:41:15.1006475Z 2025-09-07T07:41:15.1006610Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1006895Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1006975Z return mod(**inputs) 2025-09-07T07:41:15.1007342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1007435Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1007864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1007955Z layer_outputs = layer_module( 2025-09-07T07:41:15.1008300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1008397Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1008763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.1008874Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.1009240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 343, in forward 2025-09-07T07:41:15.1009408Z hidden_states = hidden_states + self.dropout(forwarded_states) 2025-09-07T07:41:15.1009413Z 2025-09-07T07:41:15.1009509Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.1009653Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1009937Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1010019Z return mod(**inputs) 2025-09-07T07:41:15.1010392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1010482Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1010850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1010936Z layer_outputs = layer_module( 2025-09-07T07:41:15.1011280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1011378Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1011744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.1011848Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.1012208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:41:15.1012350Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.1012712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.1012811Z return self.weight * hidden_states 2025-09-07T07:41:15.1012815Z 2025-09-07T07:41:15.1012950Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1013236Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1013315Z return mod(**inputs) 2025-09-07T07:41:15.1013679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1013778Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1014145Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1014231Z layer_outputs = layer_module( 2025-09-07T07:41:15.1014640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1014745Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1015115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.1015215Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.1015579Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.1015683Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.1016048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:15.1016217Z query_states = self.q(hidden_states) 2025-09-07T07:41:15.1016222Z 2025-09-07T07:41:15.1016356Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1016647Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1016724Z return mod(**inputs) 2025-09-07T07:41:15.1017090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1017184Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1017550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1017640Z layer_outputs = layer_module( 2025-09-07T07:41:15.1017980Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1018084Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1018446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.1018547Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.1018916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.1019019Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.1019385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:15.1019476Z key_states = self.k(current_states) 2025-09-07T07:41:15.1019480Z 2025-09-07T07:41:15.1019612Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1019901Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1019978Z return mod(**inputs) 2025-09-07T07:41:15.1020352Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1020442Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1020814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1020900Z layer_outputs = layer_module( 2025-09-07T07:41:15.1021243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1021343Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1021708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.1021810Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.1022174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.1022281Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.1022651Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.1022907Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.1022913Z 2025-09-07T07:41:15.1023059Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1023346Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1023421Z return mod(**inputs) 2025-09-07T07:41:15.1023791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1023879Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1024248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1024334Z layer_outputs = layer_module( 2025-09-07T07:41:15.1024743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1024840Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1025205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.1025309Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.1025673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.1025778Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.1026139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.1026308Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.1026312Z 2025-09-07T07:41:15.1026454Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1026743Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1026822Z return mod(**inputs) 2025-09-07T07:41:15.1027192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1027285Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1027651Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1027736Z layer_outputs = layer_module( 2025-09-07T07:41:15.1028082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1028178Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1028547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.1028646Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.1029013Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.1029120Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.1029487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.1029658Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.1029663Z 2025-09-07T07:41:15.1029796Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1030084Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1030161Z return mod(**inputs) 2025-09-07T07:41:15.1030529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1030622Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1030990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1031078Z layer_outputs = layer_module( 2025-09-07T07:41:15.1031502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1031606Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1031978Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.1032080Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.1032449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.1032553Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.1032915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.1033190Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.1033196Z 2025-09-07T07:41:15.1033332Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1033624Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1033700Z return mod(**inputs) 2025-09-07T07:41:15.1034070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1034160Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1034528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1034616Z layer_outputs = layer_module( 2025-09-07T07:41:15.1034956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1035057Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1035424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.1035522Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.1035893Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.1035997Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.1036363Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:15.1036458Z value_states = self.v(current_states) 2025-09-07T07:41:15.1036462Z 2025-09-07T07:41:15.1036599Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1036884Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1036961Z return mod(**inputs) 2025-09-07T07:41:15.1037333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1037423Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1037799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1037885Z layer_outputs = layer_module( 2025-09-07T07:41:15.1038227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1038325Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1038686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.1038790Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.1039155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.1039261Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.1039629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.1039766Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.1039867Z 2025-09-07T07:41:15.1040014Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1040303Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1040383Z return mod(**inputs) 2025-09-07T07:41:15.1040748Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1040838Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1041207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1041292Z layer_outputs = layer_module( 2025-09-07T07:41:15.1041713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1041810Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1042176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.1042281Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.1042643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.1042748Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.1043109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.1043245Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.1043253Z 2025-09-07T07:41:15.1043388Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1043676Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1043755Z return mod(**inputs) 2025-09-07T07:41:15.1044126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1044218Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1044584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1044667Z layer_outputs = layer_module( 2025-09-07T07:41:15.1045012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1045108Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1045473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.1045573Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.1045939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.1046046Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.1046411Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.1046551Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.1046556Z 2025-09-07T07:41:15.1046689Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1046976Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1047053Z return mod(**inputs) 2025-09-07T07:41:15.1047419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1047512Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1047876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1047970Z layer_outputs = layer_module( 2025-09-07T07:41:15.1048428Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1048538Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1048906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.1049007Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.1049381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.1049485Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.1049844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:15.1049986Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:15.1050064Z 2025-09-07T07:41:15.1050203Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1050493Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1050572Z return mod(**inputs) 2025-09-07T07:41:15.1050943Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1051032Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1051399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1051488Z layer_outputs = layer_module( 2025-09-07T07:41:15.1051831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1051929Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1052295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:41:15.1052395Z self_attention_outputs = self.layer[0]( 2025-09-07T07:41:15.1052763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:41:15.1052866Z attention_output = self.SelfAttention( 2025-09-07T07:41:15.1053232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:15.1053322Z attn_output = self.o(attn_output) 2025-09-07T07:41:15.1053327Z 2025-09-07T07:41:15.1053426Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.1053559Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1053845Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1053924Z return mod(**inputs) 2025-09-07T07:41:15.1054295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1054389Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1054758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1054844Z layer_outputs = layer_module( 2025-09-07T07:41:15.1055188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1055287Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1055652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.1055755Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.1056116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:41:15.1056261Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.1056623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.1056720Z return self.weight * hidden_states 2025-09-07T07:41:15.1056809Z 2025-09-07T07:41:15.1056953Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1057241Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1057317Z return mod(**inputs) 2025-09-07T07:41:15.1057686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1057778Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1058143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1058231Z layer_outputs = layer_module( 2025-09-07T07:41:15.1058635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1058737Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1059109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.1059212Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.1059575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.1059682Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.1060042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:41:15.1060138Z query_states = self.q(hidden_states) 2025-09-07T07:41:15.1060143Z 2025-09-07T07:41:15.1060276Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1060565Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1060647Z return mod(**inputs) 2025-09-07T07:41:15.1061015Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1061108Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1061478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1061569Z layer_outputs = layer_module( 2025-09-07T07:41:15.1061911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1062010Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1062378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.1062480Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.1062851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.1062958Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.1063329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:41:15.1063421Z key_states = self.k(current_states) 2025-09-07T07:41:15.1063425Z 2025-09-07T07:41:15.1063562Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1063847Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1063924Z return mod(**inputs) 2025-09-07T07:41:15.1064296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1064384Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1064752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1064841Z layer_outputs = layer_module( 2025-09-07T07:41:15.1065180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1065527Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1065902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.1066006Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.1066367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.1066474Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.1066840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.1067011Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.1067130Z 2025-09-07T07:41:15.1067273Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1067557Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1067639Z return mod(**inputs) 2025-09-07T07:41:15.1068008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1068097Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1068468Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1068553Z layer_outputs = layer_module( 2025-09-07T07:41:15.1068899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1068995Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1069356Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.1069465Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.1069828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.1069940Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.1070302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.1070471Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.1070480Z 2025-09-07T07:41:15.1070613Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1070899Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1070978Z return mod(**inputs) 2025-09-07T07:41:15.1071345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1071441Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1071808Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1071898Z layer_outputs = layer_module( 2025-09-07T07:41:15.1072243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1072340Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1072704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.1072806Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.1073168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.1073281Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.1073646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:41:15.1073818Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:41:15.1073823Z 2025-09-07T07:41:15.1074043Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1074341Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1074415Z return mod(**inputs) 2025-09-07T07:41:15.1074781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1074874Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1075237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1075328Z layer_outputs = layer_module( 2025-09-07T07:41:15.1075669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1075851Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1076222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.1076324Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.1076689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.1076796Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.1077156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:41:15.1077371Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:41:15.1077376Z 2025-09-07T07:41:15.1077510Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1077800Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1077880Z return mod(**inputs) 2025-09-07T07:41:15.1078253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1078346Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1078713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1078804Z layer_outputs = layer_module( 2025-09-07T07:41:15.1079147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1079246Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1079608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.1079710Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.1080078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.1080185Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.1080555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:41:15.1080649Z value_states = self.v(current_states) 2025-09-07T07:41:15.1080653Z 2025-09-07T07:41:15.1080792Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1081078Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1081154Z return mod(**inputs) 2025-09-07T07:41:15.1081530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1081621Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1081989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1082080Z layer_outputs = layer_module( 2025-09-07T07:41:15.1082426Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1082616Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1082986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.1083092Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.1083455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.1083561Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.1083928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.1084065Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.1084144Z 2025-09-07T07:41:15.1084287Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1084573Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1084656Z return mod(**inputs) 2025-09-07T07:41:15.1085024Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1085111Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1085479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1085563Z layer_outputs = layer_module( 2025-09-07T07:41:15.1085904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1085998Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1086359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.1086469Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.1086833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.1086944Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.1087305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.1087446Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.1087451Z 2025-09-07T07:41:15.1087585Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1087871Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1087952Z return mod(**inputs) 2025-09-07T07:41:15.1088319Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1088416Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1088784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1088874Z layer_outputs = layer_module( 2025-09-07T07:41:15.1089219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1089317Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1089684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.1089786Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.1090148Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.1090259Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.1090624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:41:15.1090764Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:41:15.1090770Z 2025-09-07T07:41:15.1090994Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1091298Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1091377Z return mod(**inputs) 2025-09-07T07:41:15.1091745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1091838Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1092208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1092297Z layer_outputs = layer_module( 2025-09-07T07:41:15.1092641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1092858Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1093227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.1093335Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.1093702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.1093810Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.1094178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:41:15.1094317Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:41:15.1094322Z 2025-09-07T07:41:15.1094455Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1094746Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1094826Z return mod(**inputs) 2025-09-07T07:41:15.1095196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1095289Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1095657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1095746Z layer_outputs = layer_module( 2025-09-07T07:41:15.1096088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1096187Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1096551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:41:15.1096653Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:41:15.1097019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:41:15.1097132Z attention_output = self.EncDecAttention( 2025-09-07T07:41:15.1097502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:41:15.1097595Z attn_output = self.o(attn_output) 2025-09-07T07:41:15.1097600Z 2025-09-07T07:41:15.1097700Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.1097836Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1098120Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1098201Z return mod(**inputs) 2025-09-07T07:41:15.1098570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1098664Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1099027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1099118Z layer_outputs = layer_module( 2025-09-07T07:41:15.1099461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1099650Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1100027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.1100139Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.1100501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:41:15.1100629Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:41:15.1100991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:41:15.1101085Z return self.weight * hidden_states 2025-09-07T07:41:15.1101156Z 2025-09-07T07:41:15.1101295Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1101584Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1101665Z return mod(**inputs) 2025-09-07T07:41:15.1102032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1102124Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1102489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1102578Z layer_outputs = layer_module( 2025-09-07T07:41:15.1102919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1103016Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1103383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.1103499Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.1103867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.1104021Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.1104387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:41:15.1104485Z hidden_states = self.wi(hidden_states) 2025-09-07T07:41:15.1104490Z 2025-09-07T07:41:15.1104624Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1104914Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1104992Z return mod(**inputs) 2025-09-07T07:41:15.1105361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1105453Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1105820Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1105914Z layer_outputs = layer_module( 2025-09-07T07:41:15.1106257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1106358Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1106719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.1106829Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.1107194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.1107346Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.1107718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:41:15.1107817Z hidden_states = self.act(hidden_states) 2025-09-07T07:41:15.1107822Z 2025-09-07T07:41:15.1108056Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1108352Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1108429Z return mod(**inputs) 2025-09-07T07:41:15.1108804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:41:15.1108894Z decoder_outputs = self.decoder( 2025-09-07T07:41:15.1109266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:41:15.1109353Z layer_outputs = layer_module( 2025-09-07T07:41:15.1109695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:41:15.1109896Z return super().__call__(*args, **kwargs) 2025-09-07T07:41:15.1110264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:41:15.1110378Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:41:15.1110741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:41:15.1110895Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:41:15.1111255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:41:15.1111351Z hidden_states = self.wo(hidden_states) 2025-09-07T07:41:15.1111356Z 2025-09-07T07:41:15.1111457Z cudagraph partition due to non gpu ops 2025-09-07T07:41:15.1111590Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1111884Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1111960Z return mod(**inputs) 2025-09-07T07:41:15.1112329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1789, in forward 2025-09-07T07:41:15.1112490Z sequence_output = sequence_output * (self.model_dim**-0.5) 2025-09-07T07:41:15.1112494Z 2025-09-07T07:41:15.1112626Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:41:15.1112916Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:41:15.1112994Z return mod(**inputs) 2025-09-07T07:41:15.1113361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1791, in forward 2025-09-07T07:41:15.1113464Z lm_logits = self.lm_head(sequence_output) 2025-09-07T07:41:15.1113469Z 2025-09-07T07:41:23.1938013Z pass 2025-09-07T07:41:23.1938552Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:41:25.0597607Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:41:25.0598801Z import pynvml # type: ignore[import] 2025-09-07T07:41:27.6476519Z 2025-09-07T07:41:39.4588924Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:41:39.4589232Z loading model: 0it [00:11, ?it/s] 2025-09-07T07:41:39.4844830Z cpu eval hf_T5_base 2025-09-07T07:42:14.3993511Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:42:23.2099473Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:42:32.0091664Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:43:05.9904588Z cudagraph partition due to non gpu ops 2025-09-07T07:43:05.9904936Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:05.9905857Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:05.9906353Z return mod(**inputs) 2025-09-07T07:43:05.9906892Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:05.9907463Z decoder_outputs = self.decoder( 2025-09-07T07:43:05.9908012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:05.9908573Z layer_outputs = layer_module( 2025-09-07T07:43:05.9909089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:05.9909896Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:05.9910456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:05.9911027Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:05.9911604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:05.9912174Z attention_output = self.SelfAttention( 2025-09-07T07:43:05.9912732Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 546, in forward 2025-09-07T07:43:05.9913315Z position_bias = position_bias + causal_mask 2025-09-07T07:43:05.9913526Z 2025-09-07T07:43:05.9913664Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:05.9914185Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:05.9914646Z return mod(**inputs) 2025-09-07T07:43:05.9915160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:05.9915718Z decoder_outputs = self.decoder( 2025-09-07T07:43:05.9916262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:05.9916813Z layer_outputs = layer_module( 2025-09-07T07:43:05.9917317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:05.9917854Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:05.9918409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:05.9918972Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:05.9919532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:05.9920131Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:05.9920729Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:05.9921289Z return self.weight * hidden_states 2025-09-07T07:43:05.9921475Z 2025-09-07T07:43:05.9921615Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:05.9922134Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:05.9922591Z return mod(**inputs) 2025-09-07T07:43:05.9923101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:05.9923657Z decoder_outputs = self.decoder( 2025-09-07T07:43:05.9924203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:05.9924750Z layer_outputs = layer_module( 2025-09-07T07:43:05.9925265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:05.9925803Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:05.9926493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:05.9927066Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:05.9927624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:05.9928191Z attention_output = self.SelfAttention( 2025-09-07T07:43:05.9928752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:05.9929304Z query_states = self.q(hidden_states) 2025-09-07T07:43:05.9929490Z 2025-09-07T07:43:05.9929633Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:05.9930255Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:05.9930721Z return mod(**inputs) 2025-09-07T07:43:05.9931235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:05.9931793Z decoder_outputs = self.decoder( 2025-09-07T07:43:05.9932330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:05.9932883Z layer_outputs = layer_module( 2025-09-07T07:43:05.9933393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:05.9933929Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:05.9934488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:05.9935048Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:05.9935611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:05.9936173Z attention_output = self.SelfAttention( 2025-09-07T07:43:05.9936733Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:05.9937286Z key_states = self.k(current_states) 2025-09-07T07:43:05.9937467Z 2025-09-07T07:43:05.9937939Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:05.9938461Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:05.9938926Z return mod(**inputs) 2025-09-07T07:43:05.9939436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:05.9939989Z decoder_outputs = self.decoder( 2025-09-07T07:43:05.9940535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:05.9941093Z layer_outputs = layer_module( 2025-09-07T07:43:05.9941603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:05.9942150Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:05.9942700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:05.9943263Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:05.9943823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:05.9944390Z attention_output = self.SelfAttention( 2025-09-07T07:43:05.9944956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:05.9945585Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:05.9945860Z 2025-09-07T07:43:05.9945996Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:05.9946516Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:05.9947114Z return mod(**inputs) 2025-09-07T07:43:05.9947645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:05.9948195Z decoder_outputs = self.decoder( 2025-09-07T07:43:05.9948738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:05.9949291Z layer_outputs = layer_module( 2025-09-07T07:43:05.9949798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:05.9950326Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:05.9950879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:05.9951532Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:05.9952096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:05.9952662Z attention_output = self.SelfAttention( 2025-09-07T07:43:05.9953216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:05.9953851Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:05.9954122Z 2025-09-07T07:43:05.9954258Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:05.9954775Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:05.9955238Z return mod(**inputs) 2025-09-07T07:43:05.9955742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:05.9956304Z decoder_outputs = self.decoder( 2025-09-07T07:43:05.9956841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:05.9957394Z layer_outputs = layer_module( 2025-09-07T07:43:05.9957899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:05.9958436Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:05.9958987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:05.9959548Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:05.9960107Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:05.9960671Z attention_output = self.SelfAttention( 2025-09-07T07:43:05.9961239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:05.9961873Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:05.9962138Z 2025-09-07T07:43:05.9962284Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:05.9962800Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:05.9963255Z return mod(**inputs) 2025-09-07T07:43:05.9963768Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:05.9964320Z decoder_outputs = self.decoder( 2025-09-07T07:43:05.9964861Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:05.9965704Z layer_outputs = layer_module( 2025-09-07T07:43:05.9966211Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:05.9966777Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:05.9967566Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:05.9968144Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:05.9968701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:05.9969267Z attention_output = self.SelfAttention( 2025-09-07T07:43:05.9969831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:05.9970505Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:05.9970818Z 2025-09-07T07:43:05.9970954Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:05.9971474Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:05.9972078Z return mod(**inputs) 2025-09-07T07:43:05.9972588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:05.9973148Z decoder_outputs = self.decoder( 2025-09-07T07:43:05.9973695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:05.9974244Z layer_outputs = layer_module( 2025-09-07T07:43:05.9974756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:05.9975294Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:05.9975843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:05.9976411Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:05.9976974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:05.9977542Z attention_output = self.SelfAttention( 2025-09-07T07:43:05.9978108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:05.9978667Z value_states = self.v(current_states) 2025-09-07T07:43:05.9978857Z 2025-09-07T07:43:05.9978993Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:05.9979509Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:05.9979971Z return mod(**inputs) 2025-09-07T07:43:05.9980477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:05.9981035Z decoder_outputs = self.decoder( 2025-09-07T07:43:05.9981577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:05.9982136Z layer_outputs = layer_module( 2025-09-07T07:43:05.9982646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:05.9983188Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:05.9983745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:05.9984308Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:05.9984866Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:05.9985428Z attention_output = self.SelfAttention( 2025-09-07T07:43:05.9985987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:05.9986594Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:05.9986834Z 2025-09-07T07:43:05.9986974Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:05.9987491Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:05.9988058Z return mod(**inputs) 2025-09-07T07:43:05.9988594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:05.9989152Z decoder_outputs = self.decoder( 2025-09-07T07:43:05.9989694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:05.9990246Z layer_outputs = layer_module( 2025-09-07T07:43:05.9990751Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:05.9991289Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:05.9991844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:05.9992505Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:05.9993063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:05.9993633Z attention_output = self.SelfAttention( 2025-09-07T07:43:05.9994196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:05.9994794Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:05.9995029Z 2025-09-07T07:43:05.9995170Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:05.9995683Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:05.9996149Z return mod(**inputs) 2025-09-07T07:43:05.9996665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:05.9997224Z decoder_outputs = self.decoder( 2025-09-07T07:43:05.9997768Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:05.9998321Z layer_outputs = layer_module( 2025-09-07T07:43:05.9998827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:05.9999364Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:05.9999918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0000482Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0001040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0001610Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0002170Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0002778Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0003011Z 2025-09-07T07:43:06.0003148Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0003671Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0004136Z return mod(**inputs) 2025-09-07T07:43:06.0004649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.0005205Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.0005747Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0006303Z layer_outputs = layer_module( 2025-09-07T07:43:06.0006813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0007361Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0007915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0009245Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0009841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0010407Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0010968Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.0011571Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.0011811Z 2025-09-07T07:43:06.0011947Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0012462Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0013024Z return mod(**inputs) 2025-09-07T07:43:06.0013542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.0014096Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.0014641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0015196Z layer_outputs = layer_module( 2025-09-07T07:43:06.0015708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0016247Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0016799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0017363Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0017920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0018493Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0019058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.0019617Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.0019800Z 2025-09-07T07:43:06.0019937Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0020455Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0020916Z return mod(**inputs) 2025-09-07T07:43:06.0021426Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.0021984Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.0022528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0023085Z layer_outputs = layer_module( 2025-09-07T07:43:06.0023597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0024130Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0024688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.0025256Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.0025826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.0026395Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.0026964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.0027520Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.0027710Z 2025-09-07T07:43:06.0027850Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0028373Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0028830Z return mod(**inputs) 2025-09-07T07:43:06.0029466Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0030036Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0030575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0031123Z layer_outputs = layer_module( 2025-09-07T07:43:06.0031627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0032162Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0032716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0033374Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0033928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0034499Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0035066Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.0035621Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.0035804Z 2025-09-07T07:43:06.0035947Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0036462Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0036928Z return mod(**inputs) 2025-09-07T07:43:06.0037438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0037991Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0038541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0039088Z layer_outputs = layer_module( 2025-09-07T07:43:06.0039605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0040141Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0040699Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0041257Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0041814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0042383Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0042942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.0043498Z key_states = self.k(current_states) 2025-09-07T07:43:06.0043683Z 2025-09-07T07:43:06.0043822Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0044341Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0044807Z return mod(**inputs) 2025-09-07T07:43:06.0045316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0045871Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0046410Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0046976Z layer_outputs = layer_module( 2025-09-07T07:43:06.0047491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0048030Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0048590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0049153Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0049862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0050440Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0050998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0051627Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0051903Z 2025-09-07T07:43:06.0052040Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0052553Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0053014Z return mod(**inputs) 2025-09-07T07:43:06.0053523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0054175Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0054721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0055274Z layer_outputs = layer_module( 2025-09-07T07:43:06.0055778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0056317Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0056868Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0057436Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0057995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0058565Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0059127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0059761Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0060036Z 2025-09-07T07:43:06.0060171Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0060689Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0061152Z return mod(**inputs) 2025-09-07T07:43:06.0061655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0062216Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0062760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0063313Z layer_outputs = layer_module( 2025-09-07T07:43:06.0063829Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0064363Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0064922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0065579Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0066144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0066702Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0067263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0067895Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0068163Z 2025-09-07T07:43:06.0068301Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0068823Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0069282Z return mod(**inputs) 2025-09-07T07:43:06.0069982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0070552Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0071100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0071652Z layer_outputs = layer_module( 2025-09-07T07:43:06.0072159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0072702Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0073252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0073818Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0074513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0075076Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0075644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0076313Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0076623Z 2025-09-07T07:43:06.0076762Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0077275Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0077733Z return mod(**inputs) 2025-09-07T07:43:06.0078240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0078794Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0079338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0079887Z layer_outputs = layer_module( 2025-09-07T07:43:06.0080398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0080934Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0081493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0082049Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0082606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0083166Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0083728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0084405Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0084713Z 2025-09-07T07:43:06.0084856Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0085373Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0085837Z return mod(**inputs) 2025-09-07T07:43:06.0086348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0086905Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0087447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0087997Z layer_outputs = layer_module( 2025-09-07T07:43:06.0088510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0089048Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0089613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0090175Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0090856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0091442Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0092016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.0092567Z value_states = self.v(current_states) 2025-09-07T07:43:06.0092759Z 2025-09-07T07:43:06.0092896Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0093412Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0093870Z return mod(**inputs) 2025-09-07T07:43:06.0094384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0095036Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0095586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0096139Z layer_outputs = layer_module( 2025-09-07T07:43:06.0096649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0097185Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0097740Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0098307Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0098868Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0099435Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0100004Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0100601Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0100844Z 2025-09-07T07:43:06.0100983Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0101500Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0101962Z return mod(**inputs) 2025-09-07T07:43:06.0102467Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0103023Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0103568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0104111Z layer_outputs = layer_module( 2025-09-07T07:43:06.0104627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0105158Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0105720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0106286Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0106844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0107412Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0107972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0108577Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0108813Z 2025-09-07T07:43:06.0108950Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0109476Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0109934Z return mod(**inputs) 2025-09-07T07:43:06.0110569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0111148Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0111693Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0112243Z layer_outputs = layer_module( 2025-09-07T07:43:06.0112752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0113293Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0113849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0114414Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0115093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0115654Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0116223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0116825Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0117056Z 2025-09-07T07:43:06.0117197Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0117717Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0118180Z return mod(**inputs) 2025-09-07T07:43:06.0118691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0119250Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0119795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0120344Z layer_outputs = layer_module( 2025-09-07T07:43:06.0120857Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0121396Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0121951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0122516Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0123070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0123636Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0124197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.0124801Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.0125039Z 2025-09-07T07:43:06.0125178Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0125691Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0126158Z return mod(**inputs) 2025-09-07T07:43:06.0126666Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0127225Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0127766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0128319Z layer_outputs = layer_module( 2025-09-07T07:43:06.0128832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0129369Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0129924Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0130489Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0131170Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0131746Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0132308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.0132860Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.0133038Z 2025-09-07T07:43:06.0133136Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.0133456Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0133972Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0134432Z return mod(**inputs) 2025-09-07T07:43:06.0134934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0135579Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0136123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0136675Z layer_outputs = layer_module( 2025-09-07T07:43:06.0137182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0137710Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0138267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0138836Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0139408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.0139997Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0140583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0141136Z return self.weight * hidden_states 2025-09-07T07:43:06.0141323Z 2025-09-07T07:43:06.0141462Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0141974Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0142432Z return mod(**inputs) 2025-09-07T07:43:06.0142938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0143492Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0144032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0144580Z layer_outputs = layer_module( 2025-09-07T07:43:06.0145085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0145618Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0146174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0146742Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0147311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0147927Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0148537Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.0149095Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.0149284Z 2025-09-07T07:43:06.0149422Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0149937Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0150398Z return mod(**inputs) 2025-09-07T07:43:06.0151025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0151591Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0152134Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0152680Z layer_outputs = layer_module( 2025-09-07T07:43:06.0153185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0153721Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0154280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0154847Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0155509Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0156121Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0156737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.0157295Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.0157486Z 2025-09-07T07:43:06.0157620Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0158136Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0158598Z return mod(**inputs) 2025-09-07T07:43:06.0159101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0159656Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0160195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0160748Z layer_outputs = layer_module( 2025-09-07T07:43:06.0161258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0161793Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0162339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0162912Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0163479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0164091Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0164700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.0165258Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.0165536Z 2025-09-07T07:43:06.0165637Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.0165956Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0166476Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0166933Z return mod(**inputs) 2025-09-07T07:43:06.0167442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0167995Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0168536Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0169084Z layer_outputs = layer_module( 2025-09-07T07:43:06.0169587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0170124Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0170674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0171234Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0171974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.0172585Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0173181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0173732Z return self.weight * hidden_states 2025-09-07T07:43:06.0173916Z 2025-09-07T07:43:06.0174054Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0174563Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0175022Z return mod(**inputs) 2025-09-07T07:43:06.0175651Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0176205Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0176746Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0177290Z layer_outputs = layer_module( 2025-09-07T07:43:06.0177798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0178332Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0178887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0179446Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0179995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0180563Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0181122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.0181672Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.0181860Z 2025-09-07T07:43:06.0181996Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0182511Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0182973Z return mod(**inputs) 2025-09-07T07:43:06.0183479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0184036Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0184571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0185121Z layer_outputs = layer_module( 2025-09-07T07:43:06.0185634Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0186168Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0186720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0187275Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0187833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0188395Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0188956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.0189502Z key_states = self.k(current_states) 2025-09-07T07:43:06.0189686Z 2025-09-07T07:43:06.0189823Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0190343Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0190802Z return mod(**inputs) 2025-09-07T07:43:06.0191461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0192025Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0192563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0193110Z layer_outputs = layer_module( 2025-09-07T07:43:06.0193614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0194140Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0194694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0195252Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0195918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0196478Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0197043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0197669Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0197938Z 2025-09-07T07:43:06.0198072Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0198585Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0199043Z return mod(**inputs) 2025-09-07T07:43:06.0199542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0200095Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0200636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0201190Z layer_outputs = layer_module( 2025-09-07T07:43:06.0201697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0202230Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0202781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0203342Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0203894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0204452Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0205011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0205636Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0205905Z 2025-09-07T07:43:06.0206044Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0206559Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0207017Z return mod(**inputs) 2025-09-07T07:43:06.0207524Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0208075Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0208613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0209161Z layer_outputs = layer_module( 2025-09-07T07:43:06.0209665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0210203Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0210762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0211322Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0212035Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0212607Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0213163Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0213792Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0214058Z 2025-09-07T07:43:06.0214196Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0214706Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0215165Z return mod(**inputs) 2025-09-07T07:43:06.0215670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0216317Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0216860Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0217408Z layer_outputs = layer_module( 2025-09-07T07:43:06.0217910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0218442Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0218990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0219547Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0220104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0220667Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0221229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0221900Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0222212Z 2025-09-07T07:43:06.0222347Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0222862Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0223319Z return mod(**inputs) 2025-09-07T07:43:06.0223838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0250295Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0250907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0251478Z layer_outputs = layer_module( 2025-09-07T07:43:06.0252012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0252568Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0253139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0253717Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0254281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0254853Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0255416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0256092Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0256405Z 2025-09-07T07:43:06.0256545Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0257081Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0257549Z return mod(**inputs) 2025-09-07T07:43:06.0258218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0258796Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0259341Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0259888Z layer_outputs = layer_module( 2025-09-07T07:43:06.0260389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0260926Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0261475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0262033Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0262669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0263228Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0263791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.0264346Z value_states = self.v(current_states) 2025-09-07T07:43:06.0264533Z 2025-09-07T07:43:06.0264676Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0265199Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0265747Z return mod(**inputs) 2025-09-07T07:43:06.0266256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0266814Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0267357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0267907Z layer_outputs = layer_module( 2025-09-07T07:43:06.0268419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0268954Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0269510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0270072Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0270625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0271189Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0271750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0272349Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0272588Z 2025-09-07T07:43:06.0272728Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0273242Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0273706Z return mod(**inputs) 2025-09-07T07:43:06.0274211Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0274765Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0275301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0275848Z layer_outputs = layer_module( 2025-09-07T07:43:06.0276356Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0276891Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0277450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0278006Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0278753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0279328Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0279887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0280482Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0280716Z 2025-09-07T07:43:06.0280853Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0281368Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0281829Z return mod(**inputs) 2025-09-07T07:43:06.0282331Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0283002Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0283538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0284092Z layer_outputs = layer_module( 2025-09-07T07:43:06.0284600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0285138Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0285687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0286245Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0286794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0287354Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0287911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0288506Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0288739Z 2025-09-07T07:43:06.0288880Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0289393Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0289854Z return mod(**inputs) 2025-09-07T07:43:06.0290355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0290906Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0291440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0291989Z layer_outputs = layer_module( 2025-09-07T07:43:06.0292495Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0293030Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0293582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0294144Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0294699Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0295254Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0295811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.0296411Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.0296642Z 2025-09-07T07:43:06.0296780Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0297292Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0297752Z return mod(**inputs) 2025-09-07T07:43:06.0298256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0298921Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0299474Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0300021Z layer_outputs = layer_module( 2025-09-07T07:43:06.0300523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0301059Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0301611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0302169Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0302715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0303367Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0303923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.0304474Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.0304648Z 2025-09-07T07:43:06.0304786Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0305298Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0305754Z return mod(**inputs) 2025-09-07T07:43:06.0306259Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0306811Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0307349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0307898Z layer_outputs = layer_module( 2025-09-07T07:43:06.0308401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0308935Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0309485Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0310049Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0310615Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.0311197Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0311779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0312328Z return self.weight * hidden_states 2025-09-07T07:43:06.0312515Z 2025-09-07T07:43:06.0312647Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0313155Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0313612Z return mod(**inputs) 2025-09-07T07:43:06.0314119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0314671Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0315205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0315753Z layer_outputs = layer_module( 2025-09-07T07:43:06.0316255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0316790Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0317339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0317912Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0318481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0319205Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0319830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.0320384Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.0320575Z 2025-09-07T07:43:06.0320711Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0321222Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0321679Z return mod(**inputs) 2025-09-07T07:43:06.0322182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0322848Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0323391Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0323938Z layer_outputs = layer_module( 2025-09-07T07:43:06.0324445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0324973Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0325530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0326103Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0326672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0327280Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0327887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.0328447Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.0328641Z 2025-09-07T07:43:06.0328775Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0329290Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0329744Z return mod(**inputs) 2025-09-07T07:43:06.0330251Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0330802Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0331340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0331885Z layer_outputs = layer_module( 2025-09-07T07:43:06.0332390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0332931Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0333480Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0334052Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0334622Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0335229Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0335841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.0336396Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.0336583Z 2025-09-07T07:43:06.0336685Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.0336999Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0337510Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0337974Z return mod(**inputs) 2025-09-07T07:43:06.0338483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0339157Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0339710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0340256Z layer_outputs = layer_module( 2025-09-07T07:43:06.0340761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0341294Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0341841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0342400Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0343034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.0343629Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0344225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0344773Z return self.weight * hidden_states 2025-09-07T07:43:06.0344958Z 2025-09-07T07:43:06.0345091Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0345601Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0346060Z return mod(**inputs) 2025-09-07T07:43:06.0346563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0347115Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0347668Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0348223Z layer_outputs = layer_module( 2025-09-07T07:43:06.0348728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0349269Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0349823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0350385Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0350940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0351506Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0352062Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.0352614Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.0352805Z 2025-09-07T07:43:06.0352945Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0353455Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0353911Z return mod(**inputs) 2025-09-07T07:43:06.0354418Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0354972Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0355510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0356053Z layer_outputs = layer_module( 2025-09-07T07:43:06.0356557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0357091Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0357643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0358202Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0358756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0359438Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0360016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.0360565Z key_states = self.k(current_states) 2025-09-07T07:43:06.0360749Z 2025-09-07T07:43:06.0360883Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0361396Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0361850Z return mod(**inputs) 2025-09-07T07:43:06.0362354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0362987Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0363521Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0364071Z layer_outputs = layer_module( 2025-09-07T07:43:06.0364580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0365112Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0365748Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0366308Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0366864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0367424Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0367982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0368614Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0368884Z 2025-09-07T07:43:06.0369017Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0369536Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0369994Z return mod(**inputs) 2025-09-07T07:43:06.0370495Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0371046Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0371588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0372134Z layer_outputs = layer_module( 2025-09-07T07:43:06.0372637Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0373174Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0373725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0374291Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0374848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0375410Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0375963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0376592Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0376862Z 2025-09-07T07:43:06.0376996Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0377509Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0377973Z return mod(**inputs) 2025-09-07T07:43:06.0378480Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0379037Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0379736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0380310Z layer_outputs = layer_module( 2025-09-07T07:43:06.0380828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0381365Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0381921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0382479Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0383033Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0383718Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0384275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0384906Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0385171Z 2025-09-07T07:43:06.0385308Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0385819Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0386279Z return mod(**inputs) 2025-09-07T07:43:06.0386784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0387337Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0387873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0388428Z layer_outputs = layer_module( 2025-09-07T07:43:06.0388948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0389499Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0390059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0390621Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0391176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0391738Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0392295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0392973Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0393285Z 2025-09-07T07:43:06.0393425Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0393937Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0394409Z return mod(**inputs) 2025-09-07T07:43:06.0394921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0395482Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0396026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0396582Z layer_outputs = layer_module( 2025-09-07T07:43:06.0397096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0397631Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0398188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0398782Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0399339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0400029Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0400607Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0401276Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0401583Z 2025-09-07T07:43:06.0401718Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0402231Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0402688Z return mod(**inputs) 2025-09-07T07:43:06.0403192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0403824Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0404364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0404916Z layer_outputs = layer_module( 2025-09-07T07:43:06.0405422Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0405955Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0406503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0407063Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0407616Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0408177Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0408732Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.0409287Z value_states = self.v(current_states) 2025-09-07T07:43:06.0409476Z 2025-09-07T07:43:06.0409610Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0410124Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0410584Z return mod(**inputs) 2025-09-07T07:43:06.0411087Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0411641Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0412183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0412733Z layer_outputs = layer_module( 2025-09-07T07:43:06.0413237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0413775Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0414328Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0414892Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0415446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0416005Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0416563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0417160Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0417393Z 2025-09-07T07:43:06.0417533Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0418045Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0418507Z return mod(**inputs) 2025-09-07T07:43:06.0419014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0419568Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0420226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0420788Z layer_outputs = layer_module( 2025-09-07T07:43:06.0421292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0421828Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0422381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0422939Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0423488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0424137Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0424694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0425294Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0425528Z 2025-09-07T07:43:06.0425665Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0426176Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0426633Z return mod(**inputs) 2025-09-07T07:43:06.0427138Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0427691Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0428229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0428777Z layer_outputs = layer_module( 2025-09-07T07:43:06.0429283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0429815Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0430369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0430924Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0431481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0432042Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0432603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0433203Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0433433Z 2025-09-07T07:43:06.0433571Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0434085Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0434542Z return mod(**inputs) 2025-09-07T07:43:06.0435050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0435610Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0436151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0436698Z layer_outputs = layer_module( 2025-09-07T07:43:06.0437202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0437736Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0438288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0438853Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0439404Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0440080Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0440649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.0441246Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.0441479Z 2025-09-07T07:43:06.0441616Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0442126Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0442583Z return mod(**inputs) 2025-09-07T07:43:06.0443089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0443731Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0444268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0444815Z layer_outputs = layer_module( 2025-09-07T07:43:06.0445325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0445865Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0446417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0446973Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0447529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0448088Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0448645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.0449199Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.0449380Z 2025-09-07T07:43:06.0449479Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.0449799Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0450316Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0450772Z return mod(**inputs) 2025-09-07T07:43:06.0451274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0451826Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0452362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0452907Z layer_outputs = layer_module( 2025-09-07T07:43:06.0453407Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0453948Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0454501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0455077Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0455649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.0456235Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0456819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0457367Z return self.weight * hidden_states 2025-09-07T07:43:06.0457552Z 2025-09-07T07:43:06.0457691Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0458204Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0458665Z return mod(**inputs) 2025-09-07T07:43:06.0459174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0459729Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0460382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0460955Z layer_outputs = layer_module( 2025-09-07T07:43:06.0461466Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0462000Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0462553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0463122Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0463690Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0464382Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0464994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.0465648Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.0465840Z 2025-09-07T07:43:06.0465978Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0466488Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0466947Z return mod(**inputs) 2025-09-07T07:43:06.0467455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0468008Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0468544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0469097Z layer_outputs = layer_module( 2025-09-07T07:43:06.0469603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0470136Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0470691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0471258Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0471829Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0472438Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0473049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.0473606Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.0473795Z 2025-09-07T07:43:06.0473933Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0474449Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0474903Z return mod(**inputs) 2025-09-07T07:43:06.0475413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0475964Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0476504Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0477052Z layer_outputs = layer_module( 2025-09-07T07:43:06.0477556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0478092Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0478638Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0479212Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0479778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0480601Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0481227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.0481780Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.0481971Z 2025-09-07T07:43:06.0482069Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.0482388Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0482903Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0483358Z return mod(**inputs) 2025-09-07T07:43:06.0483866Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0484553Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0485094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0485647Z layer_outputs = layer_module( 2025-09-07T07:43:06.0486152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0486687Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0487241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0487800Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0488355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.0488948Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0489546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0490095Z return self.weight * hidden_states 2025-09-07T07:43:06.0490280Z 2025-09-07T07:43:06.0490420Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0490932Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0491388Z return mod(**inputs) 2025-09-07T07:43:06.0491893Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0492447Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0492987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0493532Z layer_outputs = layer_module( 2025-09-07T07:43:06.0494037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0494578Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0495132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0495692Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0496248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0496809Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0497369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.0497923Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.0498108Z 2025-09-07T07:43:06.0498243Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0498760Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0499227Z return mod(**inputs) 2025-09-07T07:43:06.0499735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0500289Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0500938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0501504Z layer_outputs = layer_module( 2025-09-07T07:43:06.0502009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0502544Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0503097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0503655Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0504214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0504865Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0505425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.0505976Z key_states = self.k(current_states) 2025-09-07T07:43:06.0506162Z 2025-09-07T07:43:06.0506301Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0506814Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0507272Z return mod(**inputs) 2025-09-07T07:43:06.0507775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0508327Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0508866Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0509422Z layer_outputs = layer_module( 2025-09-07T07:43:06.0509929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0510457Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0511014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0511573Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0512128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0512691Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0513247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0513876Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0514146Z 2025-09-07T07:43:06.0514286Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0514799Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0515251Z return mod(**inputs) 2025-09-07T07:43:06.0515764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0516321Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0516862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0517412Z layer_outputs = layer_module( 2025-09-07T07:43:06.0517913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0518452Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0519001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0519567Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0520123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0520796Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0521365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0521994Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0522260Z 2025-09-07T07:43:06.0522400Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0522914Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0523372Z return mod(**inputs) 2025-09-07T07:43:06.0523878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0524532Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0525070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0525611Z layer_outputs = layer_module( 2025-09-07T07:43:06.0526121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0526657Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0527207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0527762Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0528314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0528874Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0529433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0530065Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0530327Z 2025-09-07T07:43:06.0530462Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0530982Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0531442Z return mod(**inputs) 2025-09-07T07:43:06.0531950Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0532506Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0533050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0533598Z layer_outputs = layer_module( 2025-09-07T07:43:06.0534101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0534642Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0535192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0535753Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0536309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0536869Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0537427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0538088Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0538400Z 2025-09-07T07:43:06.0538535Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0539047Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0539513Z return mod(**inputs) 2025-09-07T07:43:06.0540019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0540569Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0541228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0541789Z layer_outputs = layer_module( 2025-09-07T07:43:06.0542292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0542827Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0543376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0543935Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0544490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0545150Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0545705Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0546378Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0546689Z 2025-09-07T07:43:06.0546824Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0547339Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0547800Z return mod(**inputs) 2025-09-07T07:43:06.0548302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0548854Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0549391Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0549943Z layer_outputs = layer_module( 2025-09-07T07:43:06.0550444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0550979Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0551530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0552088Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0552640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0553197Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0553754Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.0554304Z value_states = self.v(current_states) 2025-09-07T07:43:06.0554492Z 2025-09-07T07:43:06.0554635Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0555149Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0555606Z return mod(**inputs) 2025-09-07T07:43:06.0556119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0556672Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0557209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0557755Z layer_outputs = layer_module( 2025-09-07T07:43:06.0558263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0558797Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0559349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0559917Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0560472Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0561156Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0561725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0562326Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0562558Z 2025-09-07T07:43:06.0562695Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0563206Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0563660Z return mod(**inputs) 2025-09-07T07:43:06.0564165Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0564793Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0565419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0565977Z layer_outputs = layer_module( 2025-09-07T07:43:06.0566488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0567025Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0567573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0568130Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0568686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0569248Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0569806Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0570409Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0570642Z 2025-09-07T07:43:06.0570776Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0571292Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0571749Z return mod(**inputs) 2025-09-07T07:43:06.0572255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0572806Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0573345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0573894Z layer_outputs = layer_module( 2025-09-07T07:43:06.0574402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0574941Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0575490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0576046Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0576601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0577160Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0577719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0578312Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0578548Z 2025-09-07T07:43:06.0578682Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0579194Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0579653Z return mod(**inputs) 2025-09-07T07:43:06.0580159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0580714Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0581431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0581996Z layer_outputs = layer_module( 2025-09-07T07:43:06.0582502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0583032Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0583584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0584145Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0584697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0585383Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0585938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.0586543Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.0586781Z 2025-09-07T07:43:06.0586916Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0587428Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0587885Z return mod(**inputs) 2025-09-07T07:43:06.0588395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0588948Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0589490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0590041Z layer_outputs = layer_module( 2025-09-07T07:43:06.0590549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0591085Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0591644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0592205Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0592756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0593314Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0593872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.0594424Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.0594601Z 2025-09-07T07:43:06.0594738Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0595256Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0595717Z return mod(**inputs) 2025-09-07T07:43:06.0596228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0596783Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0597323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0597866Z layer_outputs = layer_module( 2025-09-07T07:43:06.0598373Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0598906Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0599459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0600016Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0600575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T07:43:06.0601213Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T07:43:06.0601607Z 2025-09-07T07:43:06.0601715Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.0602034Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0602549Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0603009Z return mod(**inputs) 2025-09-07T07:43:06.0603516Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0604069Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0604607Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0605236Z layer_outputs = layer_module( 2025-09-07T07:43:06.0605739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0606273Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0606827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0607395Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0607965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.0608553Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0609134Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0609684Z return self.weight * hidden_states 2025-09-07T07:43:06.0609867Z 2025-09-07T07:43:06.0610002Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0610519Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0610975Z return mod(**inputs) 2025-09-07T07:43:06.0611485Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0612034Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0612572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0613119Z layer_outputs = layer_module( 2025-09-07T07:43:06.0613628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0614160Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0614706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0615283Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0615849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0616463Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0617075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.0617630Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.0617824Z 2025-09-07T07:43:06.0617958Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0618473Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0618931Z return mod(**inputs) 2025-09-07T07:43:06.0619436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0619991Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0620534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0621082Z layer_outputs = layer_module( 2025-09-07T07:43:06.0621700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0622245Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0622797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0623368Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0623936Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0624549Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0625156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.0625824Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.0626019Z 2025-09-07T07:43:06.0626152Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0626670Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0627127Z return mod(**inputs) 2025-09-07T07:43:06.0627633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0628185Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0628724Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0629272Z layer_outputs = layer_module( 2025-09-07T07:43:06.0629774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0630314Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0640388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0641078Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0641677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0642300Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0642914Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.0643470Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.0643659Z 2025-09-07T07:43:06.0643763Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.0644084Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0644594Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0645062Z return mod(**inputs) 2025-09-07T07:43:06.0645571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0646125Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0646665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0647215Z layer_outputs = layer_module( 2025-09-07T07:43:06.0647719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0648254Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0648804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0649362Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0649917Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.0650517Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0651273Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0651846Z return self.weight * hidden_states 2025-09-07T07:43:06.0652031Z 2025-09-07T07:43:06.0653378Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0653899Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0654356Z return mod(**inputs) 2025-09-07T07:43:06.0654864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0655416Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0655953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0656613Z layer_outputs = layer_module( 2025-09-07T07:43:06.0657115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0657651Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0658200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0658759Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0659313Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0659873Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0660429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.0660978Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.0661168Z 2025-09-07T07:43:06.0661309Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0661824Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0662283Z return mod(**inputs) 2025-09-07T07:43:06.0662788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0663339Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0663877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0664426Z layer_outputs = layer_module( 2025-09-07T07:43:06.0664928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0665540Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0666102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0666669Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0667224Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0667789Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0668342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.0668890Z key_states = self.k(current_states) 2025-09-07T07:43:06.0669075Z 2025-09-07T07:43:06.0669207Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0669719Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0670175Z return mod(**inputs) 2025-09-07T07:43:06.0670679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0671231Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0671775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0672320Z layer_outputs = layer_module( 2025-09-07T07:43:06.0673003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0673553Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0674106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0674662Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0675218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0675775Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0676332Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0677084Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0677349Z 2025-09-07T07:43:06.0677486Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0677998Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0678455Z return mod(**inputs) 2025-09-07T07:43:06.0678961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0679514Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0680048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0680592Z layer_outputs = layer_module( 2025-09-07T07:43:06.0681097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0681630Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0682181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0682738Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0683293Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0683852Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0684410Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0685035Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0685298Z 2025-09-07T07:43:06.0685432Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0685943Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0686401Z return mod(**inputs) 2025-09-07T07:43:06.0686912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0687460Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0687996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0688545Z layer_outputs = layer_module( 2025-09-07T07:43:06.0689051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0689583Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0690131Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0690687Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0691238Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0691804Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0692362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0693104Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0693383Z 2025-09-07T07:43:06.0693517Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0694027Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0694485Z return mod(**inputs) 2025-09-07T07:43:06.0694989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0695533Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0696072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0696616Z layer_outputs = layer_module( 2025-09-07T07:43:06.0697197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0697724Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0698277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0698837Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0699391Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0699950Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0700503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0701166Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0701475Z 2025-09-07T07:43:06.0701611Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0702129Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0702587Z return mod(**inputs) 2025-09-07T07:43:06.0703094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0703647Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0704185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0704729Z layer_outputs = layer_module( 2025-09-07T07:43:06.0705230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0705761Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0706312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0706879Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0707432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0707990Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0708547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0709213Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0709522Z 2025-09-07T07:43:06.0709661Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0710175Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0710632Z return mod(**inputs) 2025-09-07T07:43:06.0711141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0711701Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0712238Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0712781Z layer_outputs = layer_module( 2025-09-07T07:43:06.0713419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0713964Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0714517Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0715075Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0715625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0716183Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0716738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.0717373Z value_states = self.v(current_states) 2025-09-07T07:43:06.0717558Z 2025-09-07T07:43:06.0717696Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0718214Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0718673Z return mod(**inputs) 2025-09-07T07:43:06.0719176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0719727Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0720262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0720807Z layer_outputs = layer_module( 2025-09-07T07:43:06.0721310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0721842Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0722395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0722947Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0723504Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0724061Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0724616Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0725212Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0725445Z 2025-09-07T07:43:06.0725579Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0726093Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0726549Z return mod(**inputs) 2025-09-07T07:43:06.0727060Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0727607Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0728149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0728695Z layer_outputs = layer_module( 2025-09-07T07:43:06.0729199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0729730Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0730277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0730839Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0731393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0731958Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0732513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0733222Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0733486Z 2025-09-07T07:43:06.0733624Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0734135Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0734601Z return mod(**inputs) 2025-09-07T07:43:06.0735109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0735661Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0736195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0736739Z layer_outputs = layer_module( 2025-09-07T07:43:06.0737368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0737900Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0738459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0739017Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0739568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0740128Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0740684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0741281Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0741514Z 2025-09-07T07:43:06.0741647Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0742162Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0742614Z return mod(**inputs) 2025-09-07T07:43:06.0743126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0743677Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0744212Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0744761Z layer_outputs = layer_module( 2025-09-07T07:43:06.0745268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0745803Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0746355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0746910Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0747466Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0748025Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0748586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.0749181Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.0749418Z 2025-09-07T07:43:06.0749553Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0750067Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0750524Z return mod(**inputs) 2025-09-07T07:43:06.0751027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0751576Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0752122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0752668Z layer_outputs = layer_module( 2025-09-07T07:43:06.0753282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0753832Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0754383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0754942Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0755497Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0756055Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0756610Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.0757241Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.0757423Z 2025-09-07T07:43:06.0757520Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.0757836Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0758351Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0758806Z return mod(**inputs) 2025-09-07T07:43:06.0759315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0759407Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0759777Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0759862Z layer_outputs = layer_module( 2025-09-07T07:43:06.0760205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0760301Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0760672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0760783Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0761154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.0761283Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0761646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0761740Z return self.weight * hidden_states 2025-09-07T07:43:06.0761745Z 2025-09-07T07:43:06.0761878Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0762164Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0762242Z return mod(**inputs) 2025-09-07T07:43:06.0762614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0762706Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0763077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0763165Z layer_outputs = layer_module( 2025-09-07T07:43:06.0763505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0763601Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0763967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0764079Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0764446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0764606Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0764970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.0765177Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.0765182Z 2025-09-07T07:43:06.0765395Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0765686Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0765763Z return mod(**inputs) 2025-09-07T07:43:06.0766135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0766224Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0766591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0766679Z layer_outputs = layer_module( 2025-09-07T07:43:06.0767167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0767267Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0767634Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0767746Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0768113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0768265Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0768632Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.0768731Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.0768735Z 2025-09-07T07:43:06.0768869Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0769161Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0769237Z return mod(**inputs) 2025-09-07T07:43:06.0769611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0769700Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0770071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0770156Z layer_outputs = layer_module( 2025-09-07T07:43:06.0770496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0770596Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0770962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0771078Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0771442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0771594Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0771963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.0772058Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.0772062Z 2025-09-07T07:43:06.0772161Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.0772297Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0772583Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0772659Z return mod(**inputs) 2025-09-07T07:43:06.0773027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0773123Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0773490Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0773577Z layer_outputs = layer_module( 2025-09-07T07:43:06.0774086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0774195Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0774564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0774664Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0775029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.0775165Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0775530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0775705Z return self.weight * hidden_states 2025-09-07T07:43:06.0775710Z 2025-09-07T07:43:06.0775845Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0776137Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0776214Z return mod(**inputs) 2025-09-07T07:43:06.0776584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0776673Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0777038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0777126Z layer_outputs = layer_module( 2025-09-07T07:43:06.0777467Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0777570Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0777931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0778034Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0778403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0778505Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0778872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.0778965Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.0778970Z 2025-09-07T07:43:06.0779107Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0779394Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0779469Z return mod(**inputs) 2025-09-07T07:43:06.0779843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0779931Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0780306Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0780392Z layer_outputs = layer_module( 2025-09-07T07:43:06.0780730Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0780830Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0781194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0781297Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0781657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0781765Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0782133Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.0782315Z key_states = self.k(current_states) 2025-09-07T07:43:06.0782321Z 2025-09-07T07:43:06.0782466Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0785530Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0785616Z return mod(**inputs) 2025-09-07T07:43:06.0785999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0786090Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0786462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0786549Z layer_outputs = layer_module( 2025-09-07T07:43:06.0786981Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0787082Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0787451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0787554Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0787933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0788035Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0788399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0788570Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0788576Z 2025-09-07T07:43:06.0788713Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0789001Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0789077Z return mod(**inputs) 2025-09-07T07:43:06.0789449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0789539Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0789909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0789998Z layer_outputs = layer_module( 2025-09-07T07:43:06.0790337Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0790436Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0790799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0790901Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0791264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0791370Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0791735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0791903Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0791910Z 2025-09-07T07:43:06.0792049Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0792335Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0792413Z return mod(**inputs) 2025-09-07T07:43:06.0792780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0792869Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0793241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0793326Z layer_outputs = layer_module( 2025-09-07T07:43:06.0793763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0793867Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0794313Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0794413Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0794774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0794879Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0795241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0795453Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0795459Z 2025-09-07T07:43:06.0795593Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0795881Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0795958Z return mod(**inputs) 2025-09-07T07:43:06.0796322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0796415Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0796779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0796864Z layer_outputs = layer_module( 2025-09-07T07:43:06.0797208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0797304Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0797669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0797770Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0798137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0798242Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0798605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0798820Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0798826Z 2025-09-07T07:43:06.0798960Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0799249Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0799324Z return mod(**inputs) 2025-09-07T07:43:06.0799695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0799787Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0800157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0800246Z layer_outputs = layer_module( 2025-09-07T07:43:06.0800588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0800686Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0801048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0801148Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0801511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0801616Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0801980Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0802288Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0802294Z 2025-09-07T07:43:06.0802437Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0802811Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0802887Z return mod(**inputs) 2025-09-07T07:43:06.0803259Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0803348Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0803717Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0803851Z layer_outputs = layer_module( 2025-09-07T07:43:06.0804191Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0804292Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0804657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0804761Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0805121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0805223Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0805586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.0805680Z value_states = self.v(current_states) 2025-09-07T07:43:06.0805684Z 2025-09-07T07:43:06.0805821Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0806107Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0806182Z return mod(**inputs) 2025-09-07T07:43:06.0806554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0806642Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0807010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0807098Z layer_outputs = layer_module( 2025-09-07T07:43:06.0807438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0807533Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0807893Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0807995Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0808358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0808462Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0808828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0808966Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0808973Z 2025-09-07T07:43:06.0809110Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0809395Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0809474Z return mod(**inputs) 2025-09-07T07:43:06.0809839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0809928Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0810299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0810384Z layer_outputs = layer_module( 2025-09-07T07:43:06.0810808Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0810911Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0811276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0811416Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0811779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0811884Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0812243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0812417Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0812423Z 2025-09-07T07:43:06.0812556Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0812843Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0812924Z return mod(**inputs) 2025-09-07T07:43:06.0813292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0813386Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0813752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0813840Z layer_outputs = layer_module( 2025-09-07T07:43:06.0814179Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0814276Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0814643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0814744Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0815108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0815211Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0815574Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0815714Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0815719Z 2025-09-07T07:43:06.0815851Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0816139Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0816214Z return mod(**inputs) 2025-09-07T07:43:06.0816580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0816673Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0817042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0817130Z layer_outputs = layer_module( 2025-09-07T07:43:06.0817470Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0817571Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0817934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0818033Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0818397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0818499Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0818864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.0819003Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.0819007Z 2025-09-07T07:43:06.0819209Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0819505Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0819620Z return mod(**inputs) 2025-09-07T07:43:06.0819991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0820081Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0820450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0820536Z layer_outputs = layer_module( 2025-09-07T07:43:06.0820877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0821006Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0821373Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0821477Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0821838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0821943Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0822318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.0822410Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.0822415Z 2025-09-07T07:43:06.0822552Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0822835Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0822915Z return mod(**inputs) 2025-09-07T07:43:06.0823288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0823380Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0823749Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0823837Z layer_outputs = layer_module( 2025-09-07T07:43:06.0824180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0824275Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0824635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0824739Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0825103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T07:43:06.0825281Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T07:43:06.0825286Z 2025-09-07T07:43:06.0825386Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.0825522Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0825809Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0825889Z return mod(**inputs) 2025-09-07T07:43:06.0826260Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0826351Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0826719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0826808Z layer_outputs = layer_module( 2025-09-07T07:43:06.0827152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0827252Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0827697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0827821Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0828220Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.0828345Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0828709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0828804Z return self.weight * hidden_states 2025-09-07T07:43:06.0828809Z 2025-09-07T07:43:06.0828947Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0829261Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0829339Z return mod(**inputs) 2025-09-07T07:43:06.0829713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0829801Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0830169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0830256Z layer_outputs = layer_module( 2025-09-07T07:43:06.0830598Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0830696Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0831056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0831171Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0831535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0831689Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0832057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.0832153Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.0832161Z 2025-09-07T07:43:06.0832298Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0832582Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0832660Z return mod(**inputs) 2025-09-07T07:43:06.0833028Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0833120Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0833485Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0833575Z layer_outputs = layer_module( 2025-09-07T07:43:06.0833922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0834020Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0834385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0834499Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0834862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0835018Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0835380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.0835483Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.0835488Z 2025-09-07T07:43:06.0835623Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0835997Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0836084Z return mod(**inputs) 2025-09-07T07:43:06.0836452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0836598Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0836966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0837053Z layer_outputs = layer_module( 2025-09-07T07:43:06.0837392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0837487Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0837883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0837994Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0838364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0838516Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0838879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.0838979Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.0838984Z 2025-09-07T07:43:06.0839080Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.0839218Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0839504Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0839581Z return mod(**inputs) 2025-09-07T07:43:06.0839952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0840042Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0840416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0840501Z layer_outputs = layer_module( 2025-09-07T07:43:06.0840848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0840945Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0841307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0841411Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0841772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.0841911Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0842275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0842374Z return self.weight * hidden_states 2025-09-07T07:43:06.0842381Z 2025-09-07T07:43:06.0842517Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0842802Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0842882Z return mod(**inputs) 2025-09-07T07:43:06.0843253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0843345Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0843710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0843799Z layer_outputs = layer_module( 2025-09-07T07:43:06.0844144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0844239Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0844697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0844805Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0845213Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0845319Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0845682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.0845778Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.0845783Z 2025-09-07T07:43:06.0845918Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0846241Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0846321Z return mod(**inputs) 2025-09-07T07:43:06.0846689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0846781Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0847146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0847236Z layer_outputs = layer_module( 2025-09-07T07:43:06.0847574Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0847670Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0848038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0848138Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0848505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0848607Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0848969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.0849064Z key_states = self.k(current_states) 2025-09-07T07:43:06.0849070Z 2025-09-07T07:43:06.0849205Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0849493Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0849570Z return mod(**inputs) 2025-09-07T07:43:06.0849940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0850029Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0850395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0850486Z layer_outputs = layer_module( 2025-09-07T07:43:06.0850829Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0850929Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0851289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0851391Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0851756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0851859Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0852223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0852394Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0852399Z 2025-09-07T07:43:06.0852534Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0852904Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0852987Z return mod(**inputs) 2025-09-07T07:43:06.0853359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0853483Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0853853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0853940Z layer_outputs = layer_module( 2025-09-07T07:43:06.0854280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0854382Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0854779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0854882Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0855245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0855349Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0855718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0855886Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0855891Z 2025-09-07T07:43:06.0856029Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0856314Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0856392Z return mod(**inputs) 2025-09-07T07:43:06.0856761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0856852Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0857223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0857309Z layer_outputs = layer_module( 2025-09-07T07:43:06.0857651Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0857750Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0858110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0858213Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0858575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0858681Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0859045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0859213Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0859220Z 2025-09-07T07:43:06.0859357Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0859643Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0859724Z return mod(**inputs) 2025-09-07T07:43:06.0860089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0860184Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0860550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0860635Z layer_outputs = layer_module( 2025-09-07T07:43:06.0860980Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0861075Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0861525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0861633Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0861995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0862135Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0862496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0862709Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0862714Z 2025-09-07T07:43:06.0862851Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0863175Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0863250Z return mod(**inputs) 2025-09-07T07:43:06.0863625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0863720Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0864088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0864178Z layer_outputs = layer_module( 2025-09-07T07:43:06.0864520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0864617Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0864986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0865086Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0865591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0865695Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0866061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0866273Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0866280Z 2025-09-07T07:43:06.0866415Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0866702Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0866780Z return mod(**inputs) 2025-09-07T07:43:06.0867149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0867237Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0867605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0867694Z layer_outputs = layer_module( 2025-09-07T07:43:06.0868038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0868135Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0868499Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0868600Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0868962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0869063Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0869427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.0869525Z value_states = self.v(current_states) 2025-09-07T07:43:06.0869530Z 2025-09-07T07:43:06.0869666Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0870137Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0870224Z return mod(**inputs) 2025-09-07T07:43:06.0870596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0870750Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0871119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0871204Z layer_outputs = layer_module( 2025-09-07T07:43:06.0871544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0871643Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0872062Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0872165Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0872530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0872632Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0872999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0873138Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0873143Z 2025-09-07T07:43:06.0873280Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0873563Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0873642Z return mod(**inputs) 2025-09-07T07:43:06.0874009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0874101Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0874473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0874560Z layer_outputs = layer_module( 2025-09-07T07:43:06.0874902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0875002Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0875365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0875469Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0875832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0875936Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0876300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0876441Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0876446Z 2025-09-07T07:43:06.0876585Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0876868Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0876948Z return mod(**inputs) 2025-09-07T07:43:06.0877314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0877404Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0877771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0877857Z layer_outputs = layer_module( 2025-09-07T07:43:06.0878199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0878298Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0878753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0878861Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0879224Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0879374Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0879731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0879872Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0879877Z 2025-09-07T07:43:06.0880011Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0880301Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0880412Z return mod(**inputs) 2025-09-07T07:43:06.0880782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0880878Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0881244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0881334Z layer_outputs = layer_module( 2025-09-07T07:43:06.0881674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0881769Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0882135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0882236Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0882605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0882707Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0883074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.0883213Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.0883220Z 2025-09-07T07:43:06.0883354Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0883641Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0883718Z return mod(**inputs) 2025-09-07T07:43:06.0884086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0884176Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0884541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0884635Z layer_outputs = layer_module( 2025-09-07T07:43:06.0884974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0885079Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0885438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0885540Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0885904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0886007Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0886372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.0886464Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.0886471Z 2025-09-07T07:43:06.0886570Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.0886704Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0887070Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0887156Z return mod(**inputs) 2025-09-07T07:43:06.0887525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0887653Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0888019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0888106Z layer_outputs = layer_module( 2025-09-07T07:43:06.0888450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0888545Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0888940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0889053Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0889419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.0889550Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0889913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0890009Z return self.weight * hidden_states 2025-09-07T07:43:06.0890014Z 2025-09-07T07:43:06.0890150Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0890439Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0890514Z return mod(**inputs) 2025-09-07T07:43:06.0890880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0890974Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0891343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0891431Z layer_outputs = layer_module( 2025-09-07T07:43:06.0891770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0891869Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0892235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0892346Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0892710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0892864Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0893232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.0893329Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.0893334Z 2025-09-07T07:43:06.0893472Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0893759Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0893836Z return mod(**inputs) 2025-09-07T07:43:06.0894206Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0894294Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0894659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0894747Z layer_outputs = layer_module( 2025-09-07T07:43:06.0895088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0895188Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0895648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0895766Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0896130Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0896324Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0896688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.0896786Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.0896791Z 2025-09-07T07:43:06.0896928Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0897211Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0897318Z return mod(**inputs) 2025-09-07T07:43:06.0897688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0897779Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0898148Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0898234Z layer_outputs = layer_module( 2025-09-07T07:43:06.0898574Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0898673Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0899035Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0899150Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0899513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0899667Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0900032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.0900129Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.0900136Z 2025-09-07T07:43:06.0900234Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.0900369Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0900656Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0900731Z return mod(**inputs) 2025-09-07T07:43:06.0901095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0901186Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0901553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0901639Z layer_outputs = layer_module( 2025-09-07T07:43:06.0901985Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0902083Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0902450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0902549Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0902916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.0903059Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0903423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0903518Z return self.weight * hidden_states 2025-09-07T07:43:06.0903523Z 2025-09-07T07:43:06.0903657Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0904054Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0904139Z return mod(**inputs) 2025-09-07T07:43:06.0904513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0904640Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0905007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0905096Z layer_outputs = layer_module( 2025-09-07T07:43:06.0905436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0905534Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0905928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0906027Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0906395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0906497Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0906863Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.0906954Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.0906959Z 2025-09-07T07:43:06.0907095Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0907380Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0907457Z return mod(**inputs) 2025-09-07T07:43:06.0907825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0907917Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0908286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0908372Z layer_outputs = layer_module( 2025-09-07T07:43:06.0908715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0908817Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0909178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0909280Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0909642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0909749Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0910113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.0910203Z key_states = self.k(current_states) 2025-09-07T07:43:06.0910208Z 2025-09-07T07:43:06.0910349Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0910636Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0910717Z return mod(**inputs) 2025-09-07T07:43:06.0911083Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0911173Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0911544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0911629Z layer_outputs = layer_module( 2025-09-07T07:43:06.0911972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0912072Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0912513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0912622Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0912984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0913124Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0913484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0913656Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0913661Z 2025-09-07T07:43:06.0913796Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0914082Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0914195Z return mod(**inputs) 2025-09-07T07:43:06.0914561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0914657Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0915024Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0915110Z layer_outputs = layer_module( 2025-09-07T07:43:06.0915454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0915551Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0915914Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0916015Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0916382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0916486Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0916849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0917021Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0917029Z 2025-09-07T07:43:06.0917162Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0917449Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0917525Z return mod(**inputs) 2025-09-07T07:43:06.0917890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0917981Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0918346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0918434Z layer_outputs = layer_module( 2025-09-07T07:43:06.0918773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0918872Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0919236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0919337Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0919701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0919803Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0920168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0920338Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0920345Z 2025-09-07T07:43:06.0920479Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0920768Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0920938Z return mod(**inputs) 2025-09-07T07:43:06.0921316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0921446Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0921812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0921901Z layer_outputs = layer_module( 2025-09-07T07:43:06.0922242Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0922342Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0922706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0922854Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0923219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0923321Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0923688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0923901Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0923907Z 2025-09-07T07:43:06.0924047Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0924332Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0924408Z return mod(**inputs) 2025-09-07T07:43:06.0924779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0924870Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0925241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0925327Z layer_outputs = layer_module( 2025-09-07T07:43:06.0925667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0925769Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0926132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0926235Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0926597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0926703Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0927068Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.0927276Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.0927285Z 2025-09-07T07:43:06.0927423Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0927708Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0927788Z return mod(**inputs) 2025-09-07T07:43:06.0928157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0928245Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0928617Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0928704Z layer_outputs = layer_module( 2025-09-07T07:43:06.0929051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0929148Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0929602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0929710Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0930074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0930230Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0930592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.0930689Z value_states = self.v(current_states) 2025-09-07T07:43:06.0930695Z 2025-09-07T07:43:06.0930830Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0931115Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0931252Z return mod(**inputs) 2025-09-07T07:43:06.0931620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0931714Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0932079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0932166Z layer_outputs = layer_module( 2025-09-07T07:43:06.0932510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0932606Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0932969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0933069Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0933433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0933538Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0933900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0934041Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0934048Z 2025-09-07T07:43:06.0934184Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0934472Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0934548Z return mod(**inputs) 2025-09-07T07:43:06.0934915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0935008Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0935373Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0935465Z layer_outputs = layer_module( 2025-09-07T07:43:06.0935805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0935909Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0936275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0936377Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0936740Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0936842Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0937204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0937344Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0937351Z 2025-09-07T07:43:06.0937486Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0937777Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0937975Z return mod(**inputs) 2025-09-07T07:43:06.0938356Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0938481Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0938849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0938938Z layer_outputs = layer_module( 2025-09-07T07:43:06.0939280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0939379Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0939742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0939885Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0940248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0940351Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0940717Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.0940858Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.0940863Z 2025-09-07T07:43:06.0941000Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0941287Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0941363Z return mod(**inputs) 2025-09-07T07:43:06.0941731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0941823Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0942192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0942280Z layer_outputs = layer_module( 2025-09-07T07:43:06.0942626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0942724Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0943086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0943189Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0943550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0943655Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0944019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.0944158Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.0944163Z 2025-09-07T07:43:06.0944306Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0944593Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0944675Z return mod(**inputs) 2025-09-07T07:43:06.0945045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0945134Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0945503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0945588Z layer_outputs = layer_module( 2025-09-07T07:43:06.0945930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0946029Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0946394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0946587Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0946955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0947094Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0947456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.0947550Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.0947555Z 2025-09-07T07:43:06.0947688Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0947973Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0948078Z return mod(**inputs) 2025-09-07T07:43:06.0948448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0948541Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0948910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0949002Z layer_outputs = layer_module( 2025-09-07T07:43:06.0949342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0949438Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0949803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0949904Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0950269Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T07:43:06.0950446Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T07:43:06.0950451Z 2025-09-07T07:43:06.0950548Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.0950691Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0950975Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0951056Z return mod(**inputs) 2025-09-07T07:43:06.0951423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0951511Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0951880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0951965Z layer_outputs = layer_module( 2025-09-07T07:43:06.0952308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0952406Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0952773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0952885Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0953247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.0953376Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0953738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0953833Z return self.weight * hidden_states 2025-09-07T07:43:06.0953837Z 2025-09-07T07:43:06.0953972Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0954257Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0954341Z return mod(**inputs) 2025-09-07T07:43:06.0954709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0954883Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0955253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0955377Z layer_outputs = layer_module( 2025-09-07T07:43:06.0955720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0955817Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0956182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0956293Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0956657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0956846Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0957213Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.0957312Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.0957319Z 2025-09-07T07:43:06.0957453Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0957740Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0957817Z return mod(**inputs) 2025-09-07T07:43:06.0958183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0958276Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0958640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0958728Z layer_outputs = layer_module( 2025-09-07T07:43:06.0959068Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0959170Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0959532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0959644Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0960010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0960162Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0960525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.0960623Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.0960630Z 2025-09-07T07:43:06.0960764Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0961053Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0961132Z return mod(**inputs) 2025-09-07T07:43:06.0961502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0961594Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0961963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0962048Z layer_outputs = layer_module( 2025-09-07T07:43:06.0962388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0962487Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0962848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.0962964Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.0963405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.0963564Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.0963930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.0964076Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.0964081Z 2025-09-07T07:43:06.0964180Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.0964315Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0964600Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0964681Z return mod(**inputs) 2025-09-07T07:43:06.0965046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0965173Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0965623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0965717Z layer_outputs = layer_module( 2025-09-07T07:43:06.0966061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0966163Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0966530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0966632Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0966994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.0967128Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.0967491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.0967586Z return self.weight * hidden_states 2025-09-07T07:43:06.0967591Z 2025-09-07T07:43:06.0967731Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0968017Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0968096Z return mod(**inputs) 2025-09-07T07:43:06.0968463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0968556Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0968922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0969010Z layer_outputs = layer_module( 2025-09-07T07:43:06.0969355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0969455Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0969821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0969922Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0970287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0970393Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0970758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.0970850Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.0970855Z 2025-09-07T07:43:06.0970991Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0971280Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0971358Z return mod(**inputs) 2025-09-07T07:43:06.0971726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0971989Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0972369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0972508Z layer_outputs = layer_module( 2025-09-07T07:43:06.0972851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0972951Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0973312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0973416Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0973826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0973928Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0974300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.0974391Z key_states = self.k(current_states) 2025-09-07T07:43:06.0974396Z 2025-09-07T07:43:06.0974536Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0974821Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0974897Z return mod(**inputs) 2025-09-07T07:43:06.0975265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0975354Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0975723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0975810Z layer_outputs = layer_module( 2025-09-07T07:43:06.0976152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0976251Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0976613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0976720Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0977081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0977186Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0977546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0977714Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0977720Z 2025-09-07T07:43:06.0977857Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0978141Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0978223Z return mod(**inputs) 2025-09-07T07:43:06.0978593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0978684Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0979053Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0979138Z layer_outputs = layer_module( 2025-09-07T07:43:06.0979480Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0979576Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0979940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0980043Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0980492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0980603Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0980964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0981176Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0981181Z 2025-09-07T07:43:06.0981316Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0981603Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0981685Z return mod(**inputs) 2025-09-07T07:43:06.0982051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0982180Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.0982546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.0982637Z layer_outputs = layer_module( 2025-09-07T07:43:06.0982976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.0983074Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.0983442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.0983542Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.0983907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.0984009Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.0984380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.0984556Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.0984561Z 2025-09-07T07:43:06.0984699Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.0984990Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.0985067Z return mod(**inputs) 2025-09-07T07:43:06.0985437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.0985527Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1000949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1001099Z layer_outputs = layer_module( 2025-09-07T07:43:06.1001483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1001595Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1001977Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1002085Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1002459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1002562Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1002934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1003143Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1003149Z 2025-09-07T07:43:06.1003291Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1003582Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1003662Z return mod(**inputs) 2025-09-07T07:43:06.1004039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1004286Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1004671Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1004814Z layer_outputs = layer_module( 2025-09-07T07:43:06.1005160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1005258Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1005623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1005723Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1006088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1006227Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1006596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1006807Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1006814Z 2025-09-07T07:43:06.1006949Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1007239Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1007315Z return mod(**inputs) 2025-09-07T07:43:06.1007685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1007778Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1008145Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1008235Z layer_outputs = layer_module( 2025-09-07T07:43:06.1008576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1008679Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1009043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1009143Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1009510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1009610Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1009975Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1010069Z value_states = self.v(current_states) 2025-09-07T07:43:06.1010077Z 2025-09-07T07:43:06.1010209Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1010500Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1010575Z return mod(**inputs) 2025-09-07T07:43:06.1010951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1011043Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1011413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1011498Z layer_outputs = layer_module( 2025-09-07T07:43:06.1011842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1011943Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1012307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1012412Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1012864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1012973Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1013341Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1013516Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1013522Z 2025-09-07T07:43:06.1013659Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1013944Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1014020Z return mod(**inputs) 2025-09-07T07:43:06.1014389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1014523Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1014893Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1014983Z layer_outputs = layer_module( 2025-09-07T07:43:06.1015325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1015423Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1015784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1015887Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1016250Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1016355Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1016717Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1016856Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1016864Z 2025-09-07T07:43:06.1017001Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1017287Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1017367Z return mod(**inputs) 2025-09-07T07:43:06.1017733Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1017824Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1018189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1018273Z layer_outputs = layer_module( 2025-09-07T07:43:06.1018616Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1018716Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1019081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1019184Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1019543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1019651Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1020014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1020153Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1020158Z 2025-09-07T07:43:06.1020291Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1020581Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1020660Z return mod(**inputs) 2025-09-07T07:43:06.1021026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1021117Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1021579Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1021708Z layer_outputs = layer_module( 2025-09-07T07:43:06.1022052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1022147Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1022510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1022608Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1022972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1023114Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1023475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1023619Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1023624Z 2025-09-07T07:43:06.1023758Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1024047Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1024124Z return mod(**inputs) 2025-09-07T07:43:06.1024493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1024580Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1024944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1025034Z layer_outputs = layer_module( 2025-09-07T07:43:06.1025375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1025473Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1025841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1025942Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1026306Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1026408Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1026773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1026864Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1026869Z 2025-09-07T07:43:06.1026967Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1027107Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1027390Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1027468Z return mod(**inputs) 2025-09-07T07:43:06.1027835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1027928Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1028290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1028375Z layer_outputs = layer_module( 2025-09-07T07:43:06.1028721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1028815Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1029177Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1029289Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1029739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.1029874Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1030237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1030369Z return self.weight * hidden_states 2025-09-07T07:43:06.1030374Z 2025-09-07T07:43:06.1030508Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1030797Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1030872Z return mod(**inputs) 2025-09-07T07:43:06.1031237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1031372Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1031741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1031832Z layer_outputs = layer_module( 2025-09-07T07:43:06.1032171Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1032268Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1032633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1032743Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1033108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1033260Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1033626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.1033725Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.1033729Z 2025-09-07T07:43:06.1033866Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1034153Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1034230Z return mod(**inputs) 2025-09-07T07:43:06.1034599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1034686Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1035050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1035138Z layer_outputs = layer_module( 2025-09-07T07:43:06.1035480Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1035584Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1035944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1036058Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1036421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1036573Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1036939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.1037038Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.1037043Z 2025-09-07T07:43:06.1037180Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1037467Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1037545Z return mod(**inputs) 2025-09-07T07:43:06.1037914Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1038003Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1038482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1038605Z layer_outputs = layer_module( 2025-09-07T07:43:06.1038948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1039046Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1039409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1039523Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1039884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1040072Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1040440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.1040536Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.1040541Z 2025-09-07T07:43:06.1040644Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1040777Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1041062Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1041139Z return mod(**inputs) 2025-09-07T07:43:06.1041505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1041594Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1041959Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1042048Z layer_outputs = layer_module( 2025-09-07T07:43:06.1042394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1042489Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1042855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1042956Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1043319Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.1043453Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1043815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1043911Z return self.weight * hidden_states 2025-09-07T07:43:06.1043915Z 2025-09-07T07:43:06.1044049Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1044335Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1044414Z return mod(**inputs) 2025-09-07T07:43:06.1044782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1044872Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1045237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1045324Z layer_outputs = layer_module( 2025-09-07T07:43:06.1045666Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1045764Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1046125Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1046226Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1046681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1046794Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1047158Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1047286Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1047291Z 2025-09-07T07:43:06.1047426Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1047710Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1047785Z return mod(**inputs) 2025-09-07T07:43:06.1048152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1048275Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1048640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1048728Z layer_outputs = layer_module( 2025-09-07T07:43:06.1049068Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1049169Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1049532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1049635Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1049995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1050097Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1050464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1050555Z key_states = self.k(current_states) 2025-09-07T07:43:06.1050560Z 2025-09-07T07:43:06.1050701Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1050986Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1051065Z return mod(**inputs) 2025-09-07T07:43:06.1051430Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1051518Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1051887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1051972Z layer_outputs = layer_module( 2025-09-07T07:43:06.1052314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1052412Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1052774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1052879Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1053238Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1053343Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1053703Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1053872Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1053880Z 2025-09-07T07:43:06.1054014Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1054300Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1054382Z return mod(**inputs) 2025-09-07T07:43:06.1054750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1054841Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1055288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1055411Z layer_outputs = layer_module( 2025-09-07T07:43:06.1055754Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1055849Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1056213Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1056312Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1056674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1056819Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1057184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1057355Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1057360Z 2025-09-07T07:43:06.1057494Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1057781Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1057856Z return mod(**inputs) 2025-09-07T07:43:06.1058222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1058314Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1058681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1058771Z layer_outputs = layer_module( 2025-09-07T07:43:06.1059114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1059214Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1059582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1059681Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1060045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1060146Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1060506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1060680Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1060688Z 2025-09-07T07:43:06.1060821Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1061111Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1061186Z return mod(**inputs) 2025-09-07T07:43:06.1061558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1061649Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1062013Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1062102Z layer_outputs = layer_module( 2025-09-07T07:43:06.1062442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1062540Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1062902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1063003Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1063370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1063583Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1063957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1064207Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1064212Z 2025-09-07T07:43:06.1064348Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1064632Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1064706Z return mod(**inputs) 2025-09-07T07:43:06.1065079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1065204Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1065662Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1065751Z layer_outputs = layer_module( 2025-09-07T07:43:06.1066092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1066193Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1066555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1066659Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1067023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1067126Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1067488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1067698Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1067703Z 2025-09-07T07:43:06.1067844Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1068131Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1068211Z return mod(**inputs) 2025-09-07T07:43:06.1068577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1068665Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1069033Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1069118Z layer_outputs = layer_module( 2025-09-07T07:43:06.1069462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1069561Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1069922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1070026Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1070389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1070494Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1070853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1070949Z value_states = self.v(current_states) 2025-09-07T07:43:06.1070954Z 2025-09-07T07:43:06.1071088Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1071371Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1071449Z return mod(**inputs) 2025-09-07T07:43:06.1071814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1071905Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1072455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1072600Z layer_outputs = layer_module( 2025-09-07T07:43:06.1072948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1073044Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1073414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1073514Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1073876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1074045Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1074409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1074554Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1074559Z 2025-09-07T07:43:06.1074693Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1074985Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1075060Z return mod(**inputs) 2025-09-07T07:43:06.1075426Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1075518Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1075882Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1075972Z layer_outputs = layer_module( 2025-09-07T07:43:06.1076312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1076412Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1076775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1076877Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1077240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1077341Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1077704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1077842Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1077846Z 2025-09-07T07:43:06.1077982Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1078273Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1078348Z return mod(**inputs) 2025-09-07T07:43:06.1078720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1078809Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1079176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1079263Z layer_outputs = layer_module( 2025-09-07T07:43:06.1079602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1079702Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1080063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1080163Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1080527Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1080715Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1081088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1081276Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1081281Z 2025-09-07T07:43:06.1081419Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1081704Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1081781Z return mod(**inputs) 2025-09-07T07:43:06.1082155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1082288Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1082656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1082745Z layer_outputs = layer_module( 2025-09-07T07:43:06.1083090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1083189Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1083554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1083655Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1084020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1084122Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1084486Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1084626Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1084631Z 2025-09-07T07:43:06.1084764Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1085055Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1085132Z return mod(**inputs) 2025-09-07T07:43:06.1085503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1085594Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1085960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1086048Z layer_outputs = layer_module( 2025-09-07T07:43:06.1086390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1086492Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1086853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1086957Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1087324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1087431Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1087796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1087888Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1087892Z 2025-09-07T07:43:06.1088030Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1088314Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1088390Z return mod(**inputs) 2025-09-07T07:43:06.1088762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1088850Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1089309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1089402Z layer_outputs = layer_module( 2025-09-07T07:43:06.1089776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1089876Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1090241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1090345Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1090707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T07:43:06.1090927Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T07:43:06.1090933Z 2025-09-07T07:43:06.1091029Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1091162Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1091456Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1091530Z return mod(**inputs) 2025-09-07T07:43:06.1091903Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1091992Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1092359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1092446Z layer_outputs = layer_module( 2025-09-07T07:43:06.1092788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1092892Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1093256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1093373Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1093735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.1093861Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1094225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1094318Z return self.weight * hidden_states 2025-09-07T07:43:06.1094322Z 2025-09-07T07:43:06.1094459Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1094742Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1094821Z return mod(**inputs) 2025-09-07T07:43:06.1095189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1095278Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1095650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1095736Z layer_outputs = layer_module( 2025-09-07T07:43:06.1096077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1096176Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1096540Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1096654Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1097015Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1097172Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1097533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.1097729Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.1097735Z 2025-09-07T07:43:06.1097878Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1098199Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1098277Z return mod(**inputs) 2025-09-07T07:43:06.1098645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1098733Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1099102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1099238Z layer_outputs = layer_module( 2025-09-07T07:43:06.1099582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1099680Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1100050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1100166Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1100528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1100685Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1101046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.1101147Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.1101152Z 2025-09-07T07:43:06.1101288Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1101571Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1101650Z return mod(**inputs) 2025-09-07T07:43:06.1102018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1102109Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1102479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1102566Z layer_outputs = layer_module( 2025-09-07T07:43:06.1102910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1103007Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1103370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1103484Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1103850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1104007Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1104371Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.1104473Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.1104478Z 2025-09-07T07:43:06.1104574Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1104710Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1104994Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1105070Z return mod(**inputs) 2025-09-07T07:43:06.1105438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1105531Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1105898Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1106096Z layer_outputs = layer_module( 2025-09-07T07:43:06.1106448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1106581Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1106944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1107045Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1107406Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.1107546Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1107951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1108045Z return self.weight * hidden_states 2025-09-07T07:43:06.1108050Z 2025-09-07T07:43:06.1108191Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1108475Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1108555Z return mod(**inputs) 2025-09-07T07:43:06.1108921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1109010Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1109380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1109464Z layer_outputs = layer_module( 2025-09-07T07:43:06.1109806Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1109905Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1110269Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1110373Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1110737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1110845Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1111207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1111302Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1111307Z 2025-09-07T07:43:06.1111441Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1111726Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1111807Z return mod(**inputs) 2025-09-07T07:43:06.1112174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1112265Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1112635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1112723Z layer_outputs = layer_module( 2025-09-07T07:43:06.1113069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1113164Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1113528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1113629Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1113993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1114100Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1114463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1114645Z key_states = self.k(current_states) 2025-09-07T07:43:06.1114651Z 2025-09-07T07:43:06.1114792Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1115115Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1115191Z return mod(**inputs) 2025-09-07T07:43:06.1115558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1115650Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1116014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1116145Z layer_outputs = layer_module( 2025-09-07T07:43:06.1116487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1116586Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1116949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1117051Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1117415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1117517Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1117879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1118048Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1118053Z 2025-09-07T07:43:06.1118190Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1118476Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1118553Z return mod(**inputs) 2025-09-07T07:43:06.1118925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1119014Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1119381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1119469Z layer_outputs = layer_module( 2025-09-07T07:43:06.1119809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1119907Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1120272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1120377Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1120739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1120844Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1121211Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1121383Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1121388Z 2025-09-07T07:43:06.1121524Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1121811Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1121886Z return mod(**inputs) 2025-09-07T07:43:06.1122257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1122349Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1122718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1122803Z layer_outputs = layer_module( 2025-09-07T07:43:06.1123227Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1123331Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1123728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1123832Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1124196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1124301Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1124663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1124868Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1124874Z 2025-09-07T07:43:06.1125013Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1125302Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1125380Z return mod(**inputs) 2025-09-07T07:43:06.1125747Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1125836Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1126205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1126289Z layer_outputs = layer_module( 2025-09-07T07:43:06.1126629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1126729Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1127094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1127194Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1127560Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1127669Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1128029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1128242Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1128247Z 2025-09-07T07:43:06.1128380Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1128664Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1128745Z return mod(**inputs) 2025-09-07T07:43:06.1129112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1129203Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1129570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1129658Z layer_outputs = layer_module( 2025-09-07T07:43:06.1130001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1130097Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1130462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1130561Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1130924Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1131031Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1131392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1131691Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1131697Z 2025-09-07T07:43:06.1131836Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1132179Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1132255Z return mod(**inputs) 2025-09-07T07:43:06.1132628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1132717Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1133083Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1133212Z layer_outputs = layer_module( 2025-09-07T07:43:06.1133552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1133654Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1134016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1134119Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1134483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1134584Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1134948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1135042Z value_states = self.v(current_states) 2025-09-07T07:43:06.1135047Z 2025-09-07T07:43:06.1135184Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1135471Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1135546Z return mod(**inputs) 2025-09-07T07:43:06.1135916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1136006Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1136377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1136461Z layer_outputs = layer_module( 2025-09-07T07:43:06.1136798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1136899Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1137261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1137366Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1137729Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1137835Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1138198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1138337Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1138342Z 2025-09-07T07:43:06.1138478Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1138763Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1138839Z return mod(**inputs) 2025-09-07T07:43:06.1139204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1139294Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1139662Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1139748Z layer_outputs = layer_module( 2025-09-07T07:43:06.1140181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1140283Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1140679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1140783Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1141146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1141250Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1141612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1141786Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1141794Z 2025-09-07T07:43:06.1141928Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1142215Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1142293Z return mod(**inputs) 2025-09-07T07:43:06.1142660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1142750Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1143115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1143200Z layer_outputs = layer_module( 2025-09-07T07:43:06.1143546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1143645Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1144012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1144112Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1144476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1144583Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1144946Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1145086Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1145091Z 2025-09-07T07:43:06.1145224Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1145510Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1145585Z return mod(**inputs) 2025-09-07T07:43:06.1145954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1146045Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1146414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1146501Z layer_outputs = layer_module( 2025-09-07T07:43:06.1146843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1146939Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1147307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1147407Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1147771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1147876Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1148235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1148462Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1148467Z 2025-09-07T07:43:06.1148605Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1148924Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1149000Z return mod(**inputs) 2025-09-07T07:43:06.1149368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1149458Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1149824Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1149911Z layer_outputs = layer_module( 2025-09-07T07:43:06.1150287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1150385Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1150751Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1150853Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1151218Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1151321Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1151686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1151776Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1151781Z 2025-09-07T07:43:06.1151879Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1152015Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1152299Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1152377Z return mod(**inputs) 2025-09-07T07:43:06.1152750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1152842Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1153210Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1153294Z layer_outputs = layer_module( 2025-09-07T07:43:06.1153637Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1153734Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1154098Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1154212Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1154572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.1154703Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1155064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1155162Z return self.weight * hidden_states 2025-09-07T07:43:06.1155166Z 2025-09-07T07:43:06.1155301Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1155590Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1155667Z return mod(**inputs) 2025-09-07T07:43:06.1156034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1156128Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1156493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1156580Z layer_outputs = layer_module( 2025-09-07T07:43:06.1157006Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1157111Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1157509Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1157620Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1157983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1158135Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1158497Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.1158629Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.1158633Z 2025-09-07T07:43:06.1158766Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1159057Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1159133Z return mod(**inputs) 2025-09-07T07:43:06.1159505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1159593Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1159960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1160047Z layer_outputs = layer_module( 2025-09-07T07:43:06.1160387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1160488Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1160849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1160958Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1161326Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1161480Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1161847Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.1161945Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.1161950Z 2025-09-07T07:43:06.1162087Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1162371Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1162450Z return mod(**inputs) 2025-09-07T07:43:06.1162825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1162914Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1163283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1163370Z layer_outputs = layer_module( 2025-09-07T07:43:06.1163714Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1163815Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1164176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1164289Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1164650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1164804Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1165169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.1165574Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.1165580Z 2025-09-07T07:43:06.1165690Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1165872Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1166160Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1166236Z return mod(**inputs) 2025-09-07T07:43:06.1166601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1166692Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1167057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1167197Z layer_outputs = layer_module( 2025-09-07T07:43:06.1167537Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1167638Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1168006Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1168108Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1168472Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.1168606Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1168969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1169063Z return self.weight * hidden_states 2025-09-07T07:43:06.1169070Z 2025-09-07T07:43:06.1169203Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1169489Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1169565Z return mod(**inputs) 2025-09-07T07:43:06.1169934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1170025Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1170394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1170481Z layer_outputs = layer_module( 2025-09-07T07:43:06.1170823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1170922Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1171282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1171384Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1171747Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1171854Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1172216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1172311Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1172315Z 2025-09-07T07:43:06.1172452Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1172736Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1172812Z return mod(**inputs) 2025-09-07T07:43:06.1173182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1173274Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1173641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1173726Z layer_outputs = layer_module( 2025-09-07T07:43:06.1174152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1174259Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1174654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1174758Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1175120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1175221Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1175586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1175723Z key_states = self.k(current_states) 2025-09-07T07:43:06.1175728Z 2025-09-07T07:43:06.1175866Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1176154Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1176232Z return mod(**inputs) 2025-09-07T07:43:06.1176599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1176686Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1177056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1177140Z layer_outputs = layer_module( 2025-09-07T07:43:06.1177481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1177581Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1177942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1178045Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1178409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1178519Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1178881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1179050Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1179058Z 2025-09-07T07:43:06.1179194Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1179478Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1179556Z return mod(**inputs) 2025-09-07T07:43:06.1179925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1180017Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1180393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1180479Z layer_outputs = layer_module( 2025-09-07T07:43:06.1180826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1180922Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1181287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1181389Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1181753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1181861Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1182223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1182487Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1182493Z 2025-09-07T07:43:06.1182635Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1182958Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1183034Z return mod(**inputs) 2025-09-07T07:43:06.1183400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1183490Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1183855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1183983Z layer_outputs = layer_module( 2025-09-07T07:43:06.1184323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1184418Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1184789Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1184889Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1185254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1185357Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1185719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1185892Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1185897Z 2025-09-07T07:43:06.1186031Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1186323Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1186397Z return mod(**inputs) 2025-09-07T07:43:06.1186769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1186858Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1187226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1187315Z layer_outputs = layer_module( 2025-09-07T07:43:06.1187656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1187754Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1188116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1188218Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1188582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1188683Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1189050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1189263Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1189267Z 2025-09-07T07:43:06.1189403Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1189687Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1189763Z return mod(**inputs) 2025-09-07T07:43:06.1190132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1190222Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1190590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1190673Z layer_outputs = layer_module( 2025-09-07T07:43:06.1191089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1191193Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1191588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1191692Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1192054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1192160Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1192521Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1192799Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1192804Z 2025-09-07T07:43:06.1192942Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1193230Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1193309Z return mod(**inputs) 2025-09-07T07:43:06.1193676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1193765Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1194134Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1194220Z layer_outputs = layer_module( 2025-09-07T07:43:06.1194562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1194660Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1195022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1195132Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1195495Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1195600Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1195960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1196055Z value_states = self.v(current_states) 2025-09-07T07:43:06.1196059Z 2025-09-07T07:43:06.1196192Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1196475Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1196554Z return mod(**inputs) 2025-09-07T07:43:06.1196922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1197013Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1197382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1197467Z layer_outputs = layer_module( 2025-09-07T07:43:06.1197809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1197904Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1198268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1198367Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1198728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1198832Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1199192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1199449Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1199456Z 2025-09-07T07:43:06.1199594Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1199921Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1199998Z return mod(**inputs) 2025-09-07T07:43:06.1200362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1200452Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1200819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1200906Z layer_outputs = layer_module( 2025-09-07T07:43:06.1201293Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1201390Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1201758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1201857Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1202223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1202328Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1202691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1202829Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1202834Z 2025-09-07T07:43:06.1202968Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1203259Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1203335Z return mod(**inputs) 2025-09-07T07:43:06.1203706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1203796Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1204161Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1204249Z layer_outputs = layer_module( 2025-09-07T07:43:06.1204588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1204688Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1205048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1205152Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1205513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1205615Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1205983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1206122Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1206127Z 2025-09-07T07:43:06.1206263Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1206547Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1206622Z return mod(**inputs) 2025-09-07T07:43:06.1206991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1207080Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1207451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1207535Z layer_outputs = layer_module( 2025-09-07T07:43:06.1207969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1208074Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1208475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1208578Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1208940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1209045Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1209405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1209592Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1209596Z 2025-09-07T07:43:06.1209734Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1210024Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1210105Z return mod(**inputs) 2025-09-07T07:43:06.1210470Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1210560Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1210927Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1211012Z layer_outputs = layer_module( 2025-09-07T07:43:06.1211354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1211450Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1211817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1211919Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1212281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1212386Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1212750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1212842Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1212847Z 2025-09-07T07:43:06.1212982Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1213266Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1213345Z return mod(**inputs) 2025-09-07T07:43:06.1213709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1213803Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1214173Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1214257Z layer_outputs = layer_module( 2025-09-07T07:43:06.1214602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1214702Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1215067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1215168Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1215532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T07:43:06.1215708Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T07:43:06.1215716Z 2025-09-07T07:43:06.1215813Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1215952Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1216332Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1216417Z return mod(**inputs) 2025-09-07T07:43:06.1216786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1216920Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1217287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1217372Z layer_outputs = layer_module( 2025-09-07T07:43:06.1217715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1217810Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1218198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1218312Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1218676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.1218805Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1219171Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1219266Z return self.weight * hidden_states 2025-09-07T07:43:06.1219271Z 2025-09-07T07:43:06.1219405Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1219688Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1219765Z return mod(**inputs) 2025-09-07T07:43:06.1220132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1220225Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1220592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1220676Z layer_outputs = layer_module( 2025-09-07T07:43:06.1221019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1221118Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1221484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1221596Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1221961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1222114Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1222474Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.1222576Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.1222581Z 2025-09-07T07:43:06.1222715Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1223006Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1223081Z return mod(**inputs) 2025-09-07T07:43:06.1223448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1223539Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1223905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1223993Z layer_outputs = layer_module( 2025-09-07T07:43:06.1224335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1224431Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1224891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1225011Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1225412Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1225566Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1225932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.1226029Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.1226035Z 2025-09-07T07:43:06.1226168Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1226511Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1226588Z return mod(**inputs) 2025-09-07T07:43:06.1226960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1227048Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1227415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1227505Z layer_outputs = layer_module( 2025-09-07T07:43:06.1227846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1227946Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1228309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1228424Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1228784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1228935Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1229307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.1229407Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.1229412Z 2025-09-07T07:43:06.1229512Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1229646Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1229930Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1230010Z return mod(**inputs) 2025-09-07T07:43:06.1230376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T07:43:06.1230468Z encoder_outputs = self.encoder( 2025-09-07T07:43:06.1230833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1128, in forward 2025-09-07T07:43:06.1230968Z hidden_states = self.final_layer_norm(hidden_states) 2025-09-07T07:43:06.1231331Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1231426Z return self.weight * hidden_states 2025-09-07T07:43:06.1231430Z 2025-09-07T07:43:06.1231567Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1231853Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1231933Z return mod(**inputs) 2025-09-07T07:43:06.1232300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1232389Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1232760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1232846Z layer_outputs = layer_module( 2025-09-07T07:43:06.1233280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1233386Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1233783Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1233890Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1234251Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1234362Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1234728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1234853Z key_states = self.k(current_states) 2025-09-07T07:43:06.1234861Z 2025-09-07T07:43:06.1234994Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1235282Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1235361Z return mod(**inputs) 2025-09-07T07:43:06.1235726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1235819Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1236183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1236268Z layer_outputs = layer_module( 2025-09-07T07:43:06.1236613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1236710Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1237077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1237179Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1237541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1237652Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1238016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1238187Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1238192Z 2025-09-07T07:43:06.1238326Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1238613Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1238689Z return mod(**inputs) 2025-09-07T07:43:06.1239056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1239147Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1239516Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1239604Z layer_outputs = layer_module( 2025-09-07T07:43:06.1239945Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1240044Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1240407Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1240510Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1240874Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1240984Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1241346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1241518Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1241608Z 2025-09-07T07:43:06.1241752Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1242044Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1242154Z return mod(**inputs) 2025-09-07T07:43:06.1242525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1242614Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1242979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1243069Z layer_outputs = layer_module( 2025-09-07T07:43:06.1243448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1243555Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1243921Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1244024Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1244387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1244495Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1244857Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1245024Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1245029Z 2025-09-07T07:43:06.1245167Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1245455Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1245529Z return mod(**inputs) 2025-09-07T07:43:06.1245905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1245994Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1246362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1246450Z layer_outputs = layer_module( 2025-09-07T07:43:06.1246790Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1246889Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1247251Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1247356Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1247718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1247825Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1248195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1248402Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1248409Z 2025-09-07T07:43:06.1248548Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1248834Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1248910Z return mod(**inputs) 2025-09-07T07:43:06.1249276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1249364Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1249737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1249822Z layer_outputs = layer_module( 2025-09-07T07:43:06.1250259Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1250361Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1250761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1250866Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1251229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1251338Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1251699Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1251834Z value_states = self.v(current_states) 2025-09-07T07:43:06.1251839Z 2025-09-07T07:43:06.1251974Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1252261Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1252341Z return mod(**inputs) 2025-09-07T07:43:06.1252707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1252800Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1253166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1253250Z layer_outputs = layer_module( 2025-09-07T07:43:06.1253592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1253689Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1254055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1254157Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1254520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1254629Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1254992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1255133Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1255138Z 2025-09-07T07:43:06.1255273Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1255562Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1255638Z return mod(**inputs) 2025-09-07T07:43:06.1256004Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1256096Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1256463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1256551Z layer_outputs = layer_module( 2025-09-07T07:43:06.1256890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1256988Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1257351Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1257452Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1257817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1257923Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1258291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1258426Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1258518Z 2025-09-07T07:43:06.1258656Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1258945Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1259072Z return mod(**inputs) 2025-09-07T07:43:06.1259440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1259529Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1259894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1259981Z layer_outputs = layer_module( 2025-09-07T07:43:06.1260356Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1260454Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1260818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1260921Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1261287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1261393Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1261756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1261892Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1261897Z 2025-09-07T07:43:06.1262035Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1262321Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1262396Z return mod(**inputs) 2025-09-07T07:43:06.1262770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1262859Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1263225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1263312Z layer_outputs = layer_module( 2025-09-07T07:43:06.1263651Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1263752Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1264113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1264218Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1264583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1264693Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1265058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1265197Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1265203Z 2025-09-07T07:43:06.1265441Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1265728Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1265806Z return mod(**inputs) 2025-09-07T07:43:06.1266170Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1266259Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1266628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1266717Z layer_outputs = layer_module( 2025-09-07T07:43:06.1267203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1267305Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1267667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1267818Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1268182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1268292Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1268654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1268746Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1268800Z 2025-09-07T07:43:06.1268898Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1269033Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1269323Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1269400Z return mod(**inputs) 2025-09-07T07:43:06.1269767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1269859Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1270223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1270312Z layer_outputs = layer_module( 2025-09-07T07:43:06.1270654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1270754Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1271121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1271232Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1271600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.1271727Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1272098Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1272190Z return self.weight * hidden_states 2025-09-07T07:43:06.1272195Z 2025-09-07T07:43:06.1272332Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1272616Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1272691Z return mod(**inputs) 2025-09-07T07:43:06.1273062Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1273150Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1273523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1273608Z layer_outputs = layer_module( 2025-09-07T07:43:06.1273947Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1274048Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1274410Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1274523Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1274885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1275044Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1275408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.1275503Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.1275602Z 2025-09-07T07:43:06.1275746Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1276032Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1276148Z return mod(**inputs) 2025-09-07T07:43:06.1276514Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1276604Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1276972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1277055Z layer_outputs = layer_module( 2025-09-07T07:43:06.1277431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1277527Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1277892Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1278008Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1278372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1278526Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1278888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.1278989Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.1278993Z 2025-09-07T07:43:06.1279125Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1279412Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1279491Z return mod(**inputs) 2025-09-07T07:43:06.1279861Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1279952Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1280315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1280402Z layer_outputs = layer_module( 2025-09-07T07:43:06.1280744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1280840Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1281205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1281316Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1281681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1281832Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1282196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.1282299Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.1282304Z 2025-09-07T07:43:06.1282437Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1282724Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1282799Z return mod(**inputs) 2025-09-07T07:43:06.1283165Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1283258Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1283623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1283712Z layer_outputs = layer_module( 2025-09-07T07:43:06.1284147Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1284255Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1284619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1284759Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1285126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.1285263Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1285627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1285761Z return self.weight * hidden_states 2025-09-07T07:43:06.1285767Z 2025-09-07T07:43:06.1285902Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1286193Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1286271Z return mod(**inputs) 2025-09-07T07:43:06.1286639Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1286729Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1287094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1287182Z layer_outputs = layer_module( 2025-09-07T07:43:06.1287523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1287620Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1287985Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1288088Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1288456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1288560Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1288927Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1289018Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1289023Z 2025-09-07T07:43:06.1289159Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1289446Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1289524Z return mod(**inputs) 2025-09-07T07:43:06.1289894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1289986Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1290354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1290442Z layer_outputs = layer_module( 2025-09-07T07:43:06.1290784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1290886Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1291247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1291351Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1291716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1291821Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1292187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1292278Z key_states = self.k(current_states) 2025-09-07T07:43:06.1292283Z 2025-09-07T07:43:06.1292531Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1292825Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1292939Z return mod(**inputs) 2025-09-07T07:43:06.1293306Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1293394Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1293763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1293847Z layer_outputs = layer_module( 2025-09-07T07:43:06.1294195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1294327Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1294691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1294796Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1295156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1295263Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1295624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1295796Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1295801Z 2025-09-07T07:43:06.1295937Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1296221Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1296302Z return mod(**inputs) 2025-09-07T07:43:06.1296670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1296764Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1297130Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1297219Z layer_outputs = layer_module( 2025-09-07T07:43:06.1297563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1297659Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1298024Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1298124Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1298488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1298593Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1298956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1299128Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1299134Z 2025-09-07T07:43:06.1299268Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1299555Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1299630Z return mod(**inputs) 2025-09-07T07:43:06.1299995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1300086Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1300453Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1300543Z layer_outputs = layer_module( 2025-09-07T07:43:06.1300885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1301071Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1301443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1301578Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1301945Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1302048Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1302413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1302582Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1302619Z 2025-09-07T07:43:06.1302755Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1303044Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1303123Z return mod(**inputs) 2025-09-07T07:43:06.1303491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1303582Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1303948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1304034Z layer_outputs = layer_module( 2025-09-07T07:43:06.1304376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1304475Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1304840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1304947Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1305312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1305415Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1305782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1305994Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1305998Z 2025-09-07T07:43:06.1306135Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1306419Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1306496Z return mod(**inputs) 2025-09-07T07:43:06.1306866Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1306956Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1307323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1307411Z layer_outputs = layer_module( 2025-09-07T07:43:06.1307755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1307853Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1308215Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1308317Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1308677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1308783Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1309149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1309242Z value_states = self.v(current_states) 2025-09-07T07:43:06.1309246Z 2025-09-07T07:43:06.1309471Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1309760Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1309874Z return mod(**inputs) 2025-09-07T07:43:06.1310241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1310329Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1310696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1310782Z layer_outputs = layer_module( 2025-09-07T07:43:06.1311125Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1311280Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1311646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1311749Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1312110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1312219Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1312580Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1312720Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1312725Z 2025-09-07T07:43:06.1312859Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1313144Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1313225Z return mod(**inputs) 2025-09-07T07:43:06.1313591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1313681Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1314050Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1314141Z layer_outputs = layer_module( 2025-09-07T07:43:06.1314479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1314575Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1314939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1315037Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1315402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1315505Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1315870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1316013Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1316018Z 2025-09-07T07:43:06.1316155Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1316446Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1316522Z return mod(**inputs) 2025-09-07T07:43:06.1316887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1316979Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1317343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1317436Z layer_outputs = layer_module( 2025-09-07T07:43:06.1317777Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1317949Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1318316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1318438Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1318803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1318905Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1319270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1319406Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1319432Z 2025-09-07T07:43:06.1319568Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1319854Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1319929Z return mod(**inputs) 2025-09-07T07:43:06.1320301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1320391Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1320764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1320849Z layer_outputs = layer_module( 2025-09-07T07:43:06.1321191Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1321289Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1321652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1321758Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1322120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1322226Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1322591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1322732Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1322737Z 2025-09-07T07:43:06.1322873Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1323159Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1323235Z return mod(**inputs) 2025-09-07T07:43:06.1323603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1323694Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1324063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1324148Z layer_outputs = layer_module( 2025-09-07T07:43:06.1324494Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1324602Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1329273Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1329441Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1329829Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1329936Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1330300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1330398Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1330404Z 2025-09-07T07:43:06.1330503Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1330809Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1331107Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1331263Z return mod(**inputs) 2025-09-07T07:43:06.1331631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1331725Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1332093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1332182Z layer_outputs = layer_module( 2025-09-07T07:43:06.1332522Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1332671Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1333034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1333142Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1333501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:43:06.1333640Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1334001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1334094Z return self.weight * hidden_states 2025-09-07T07:43:06.1334099Z 2025-09-07T07:43:06.1334235Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1334518Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1334597Z return mod(**inputs) 2025-09-07T07:43:06.1334964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1335051Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1335421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1335509Z layer_outputs = layer_module( 2025-09-07T07:43:06.1335853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1335948Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1336313Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1336417Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1336776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1336889Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1337252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1337347Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1337352Z 2025-09-07T07:43:06.1337485Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1337772Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1337850Z return mod(**inputs) 2025-09-07T07:43:06.1338217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1338307Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1338672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1338759Z layer_outputs = layer_module( 2025-09-07T07:43:06.1339100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1339197Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1339656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1339798Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1340165Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1340272Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1340634Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1340728Z key_states = self.k(current_states) 2025-09-07T07:43:06.1340733Z 2025-09-07T07:43:06.1340868Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1341196Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1341271Z return mod(**inputs) 2025-09-07T07:43:06.1341639Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1341730Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1342094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1342180Z layer_outputs = layer_module( 2025-09-07T07:43:06.1342518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1342613Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1342976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1343079Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1343445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1343555Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1343919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1344089Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1344094Z 2025-09-07T07:43:06.1344229Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1344517Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1344592Z return mod(**inputs) 2025-09-07T07:43:06.1344960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1345049Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1345414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1345502Z layer_outputs = layer_module( 2025-09-07T07:43:06.1345844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1345942Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1346305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1346409Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1346770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1346877Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1347241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1347411Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1347416Z 2025-09-07T07:43:06.1347552Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1347935Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1348015Z return mod(**inputs) 2025-09-07T07:43:06.1348419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1348507Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1348872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1348958Z layer_outputs = layer_module( 2025-09-07T07:43:06.1349301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1349445Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1349810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1349914Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1350275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1350386Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1350747Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1350913Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1350918Z 2025-09-07T07:43:06.1351053Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1351339Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1351418Z return mod(**inputs) 2025-09-07T07:43:06.1351784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1351873Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1352246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1352330Z layer_outputs = layer_module( 2025-09-07T07:43:06.1352674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1352770Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1353131Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1353233Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1353593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1353704Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1354065Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1354282Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1354287Z 2025-09-07T07:43:06.1354422Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1354710Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1354789Z return mod(**inputs) 2025-09-07T07:43:06.1355157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1355249Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1355613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1355702Z layer_outputs = layer_module( 2025-09-07T07:43:06.1356043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1356138Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1356588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1356760Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1357123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1357230Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1357589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1357683Z value_states = self.v(current_states) 2025-09-07T07:43:06.1357688Z 2025-09-07T07:43:06.1357823Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1358180Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1358255Z return mod(**inputs) 2025-09-07T07:43:06.1358623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1358715Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1359080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1359167Z layer_outputs = layer_module( 2025-09-07T07:43:06.1359507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1359607Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1359967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1360071Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1360434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1360543Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1360907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1361046Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1361051Z 2025-09-07T07:43:06.1361184Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1361472Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1361548Z return mod(**inputs) 2025-09-07T07:43:06.1361915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1362003Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1362373Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1362459Z layer_outputs = layer_module( 2025-09-07T07:43:06.1362801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1362901Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1363263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1363366Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1363725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1363830Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1364197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1364337Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1364342Z 2025-09-07T07:43:06.1364479Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1364852Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1364932Z return mod(**inputs) 2025-09-07T07:43:06.1365445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1365535Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1365904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1365988Z layer_outputs = layer_module( 2025-09-07T07:43:06.1366330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1366523Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1366884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1366990Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1367351Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1367464Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1367825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1367962Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1367968Z 2025-09-07T07:43:06.1368104Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1368387Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1368464Z return mod(**inputs) 2025-09-07T07:43:06.1368831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1368921Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1369286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1369370Z layer_outputs = layer_module( 2025-09-07T07:43:06.1369714Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1369809Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1370174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1370275Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1370633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1370746Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1371106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1371250Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1371255Z 2025-09-07T07:43:06.1371389Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1371676Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1371755Z return mod(**inputs) 2025-09-07T07:43:06.1372121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1372212Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1372576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1372665Z layer_outputs = layer_module( 2025-09-07T07:43:06.1373006Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1373102Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1373601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1373711Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1374129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1374235Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1374597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1374690Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1374694Z 2025-09-07T07:43:06.1374791Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1374971Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1375254Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1375330Z return mod(**inputs) 2025-09-07T07:43:06.1375701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1375790Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1376162Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1376246Z layer_outputs = layer_module( 2025-09-07T07:43:06.1376586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1376684Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1377045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1377160Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1377525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.1377655Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1378019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1378113Z return self.weight * hidden_states 2025-09-07T07:43:06.1378118Z 2025-09-07T07:43:06.1378252Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1378534Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1378610Z return mod(**inputs) 2025-09-07T07:43:06.1378974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1379065Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1379430Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1379513Z layer_outputs = layer_module( 2025-09-07T07:43:06.1379861Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1379957Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1380324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1380435Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1380796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1380951Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1381311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.1381413Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.1381418Z 2025-09-07T07:43:06.1381551Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1381925Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1382005Z return mod(**inputs) 2025-09-07T07:43:06.1382407Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1382497Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1382861Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1382947Z layer_outputs = layer_module( 2025-09-07T07:43:06.1383288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1383415Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1383778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1383889Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1384256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1384409Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1384771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.1384872Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.1384876Z 2025-09-07T07:43:06.1385008Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1385292Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1385370Z return mod(**inputs) 2025-09-07T07:43:06.1385738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1385827Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1386194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1386282Z layer_outputs = layer_module( 2025-09-07T07:43:06.1386622Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1386719Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1387080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1387190Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1387553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1387706Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1388068Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.1388167Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.1388172Z 2025-09-07T07:43:06.1388270Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1388404Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1388687Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1388766Z return mod(**inputs) 2025-09-07T07:43:06.1389129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1389221Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1389585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1389671Z layer_outputs = layer_module( 2025-09-07T07:43:06.1390013Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1390193Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1390562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1390709Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1391071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.1391208Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1391568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1391665Z return self.weight * hidden_states 2025-09-07T07:43:06.1391701Z 2025-09-07T07:43:06.1391840Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1392128Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1392204Z return mod(**inputs) 2025-09-07T07:43:06.1392576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1392669Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1393033Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1393120Z layer_outputs = layer_module( 2025-09-07T07:43:06.1393460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1393555Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1393918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1394019Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1394379Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1394485Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1394846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1394945Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1394950Z 2025-09-07T07:43:06.1395084Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1395370Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1395445Z return mod(**inputs) 2025-09-07T07:43:06.1395814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1395903Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1396266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1396353Z layer_outputs = layer_module( 2025-09-07T07:43:06.1396696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1396797Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1397158Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1397257Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1397622Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1397723Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1398086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1398176Z key_states = self.k(current_states) 2025-09-07T07:43:06.1398181Z 2025-09-07T07:43:06.1398316Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1398668Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1398750Z return mod(**inputs) 2025-09-07T07:43:06.1399149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1399239Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1399603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1399687Z layer_outputs = layer_module( 2025-09-07T07:43:06.1400026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1400154Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1400515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1400617Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1400979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1401082Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1401443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1401610Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1401615Z 2025-09-07T07:43:06.1401751Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1402035Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1402112Z return mod(**inputs) 2025-09-07T07:43:06.1402479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1402567Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1402936Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1403022Z layer_outputs = layer_module( 2025-09-07T07:43:06.1403365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1403461Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1403823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1403927Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1404288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1404395Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1404755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1404925Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1404933Z 2025-09-07T07:43:06.1405066Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1405351Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1405430Z return mod(**inputs) 2025-09-07T07:43:06.1405795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1405886Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1406249Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1406336Z layer_outputs = layer_module( 2025-09-07T07:43:06.1406679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1406773Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1407221Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1407326Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1407718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1407823Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1408183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1408351Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1408356Z 2025-09-07T07:43:06.1408489Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1408819Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1408895Z return mod(**inputs) 2025-09-07T07:43:06.1409264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1409356Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1409721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1409817Z layer_outputs = layer_module( 2025-09-07T07:43:06.1410160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1410257Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1410619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1410722Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1411082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1411187Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1411550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1411766Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1411771Z 2025-09-07T07:43:06.1411904Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1412190Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1412268Z return mod(**inputs) 2025-09-07T07:43:06.1412635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1412729Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1413094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1413179Z layer_outputs = layer_module( 2025-09-07T07:43:06.1413525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1413622Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1413989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1414088Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1414452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1414554Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1414916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1415015Z value_states = self.v(current_states) 2025-09-07T07:43:06.1415019Z 2025-09-07T07:43:06.1415152Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1415528Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1415610Z return mod(**inputs) 2025-09-07T07:43:06.1415978Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1416104Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1416469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1416556Z layer_outputs = layer_module( 2025-09-07T07:43:06.1416895Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1417020Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1417382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1417481Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1417848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1417952Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1418316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1418456Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1418460Z 2025-09-07T07:43:06.1418595Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1418884Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1418960Z return mod(**inputs) 2025-09-07T07:43:06.1419330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1419418Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1419788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1419874Z layer_outputs = layer_module( 2025-09-07T07:43:06.1420217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1420320Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1420683Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1420783Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1421146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1421251Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1421616Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1421756Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1421761Z 2025-09-07T07:43:06.1421898Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1422184Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1422261Z return mod(**inputs) 2025-09-07T07:43:06.1422629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1422719Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1423088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1423174Z layer_outputs = layer_module( 2025-09-07T07:43:06.1423514Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1423613Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1424077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1424185Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1424583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1424687Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1425048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1425186Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1425191Z 2025-09-07T07:43:06.1425327Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1425651Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1425728Z return mod(**inputs) 2025-09-07T07:43:06.1426098Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1426186Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1426554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1426641Z layer_outputs = layer_module( 2025-09-07T07:43:06.1426984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1427079Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1427442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1427544Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1427908Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1428013Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1428378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1428519Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1428526Z 2025-09-07T07:43:06.1428659Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1428944Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1429022Z return mod(**inputs) 2025-09-07T07:43:06.1429388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1429479Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1429845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1429929Z layer_outputs = layer_module( 2025-09-07T07:43:06.1430275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1430371Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1430738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1430841Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1431206Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1431309Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1431669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1431763Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1431768Z 2025-09-07T07:43:06.1431902Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1432190Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1432350Z return mod(**inputs) 2025-09-07T07:43:06.1432721Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1432834Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1433200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1433288Z layer_outputs = layer_module( 2025-09-07T07:43:06.1433628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1433723Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1434111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1434213Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1434582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T07:43:06.1434760Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T07:43:06.1434766Z 2025-09-07T07:43:06.1434866Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1435000Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1435284Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1435362Z return mod(**inputs) 2025-09-07T07:43:06.1435729Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1435819Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1436187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1436272Z layer_outputs = layer_module( 2025-09-07T07:43:06.1436620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1436717Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1437081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1437183Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1437543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:43:06.1437682Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1438042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1438141Z return self.weight * hidden_states 2025-09-07T07:43:06.1438146Z 2025-09-07T07:43:06.1438281Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1438571Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1438647Z return mod(**inputs) 2025-09-07T07:43:06.1439014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1439108Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1439472Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1439560Z layer_outputs = layer_module( 2025-09-07T07:43:06.1439898Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1439993Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1440361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1440463Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1440911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1441027Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1441423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1441514Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1441519Z 2025-09-07T07:43:06.1441655Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1441943Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1442017Z return mod(**inputs) 2025-09-07T07:43:06.1442387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1442516Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1442886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1442974Z layer_outputs = layer_module( 2025-09-07T07:43:06.1443313Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1443414Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1443776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1443878Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1444244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1444350Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1444716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1444805Z key_states = self.k(current_states) 2025-09-07T07:43:06.1444810Z 2025-09-07T07:43:06.1444949Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1445236Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1445315Z return mod(**inputs) 2025-09-07T07:43:06.1445683Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1445773Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1446140Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1446224Z layer_outputs = layer_module( 2025-09-07T07:43:06.1446565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1446668Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1447032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1447137Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1447500Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1447613Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1447977Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1448147Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1448152Z 2025-09-07T07:43:06.1448288Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1448575Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1448654Z return mod(**inputs) 2025-09-07T07:43:06.1449022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1449192Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1449567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1449684Z layer_outputs = layer_module( 2025-09-07T07:43:06.1450027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1450123Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1450489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1450593Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1450989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1451098Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1451463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1451635Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1451642Z 2025-09-07T07:43:06.1451775Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1452058Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1452137Z return mod(**inputs) 2025-09-07T07:43:06.1452501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1452593Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1452958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1453046Z layer_outputs = layer_module( 2025-09-07T07:43:06.1453392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1453487Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1453852Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1453956Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1454322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1454428Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1454788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1454959Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1454966Z 2025-09-07T07:43:06.1455099Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1455387Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1455465Z return mod(**inputs) 2025-09-07T07:43:06.1455830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1455923Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1456287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1456376Z layer_outputs = layer_module( 2025-09-07T07:43:06.1456718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1456813Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1457184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1457289Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1457758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1457874Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1458271Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1458481Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1458486Z 2025-09-07T07:43:06.1458619Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1458912Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1458990Z return mod(**inputs) 2025-09-07T07:43:06.1459360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1459492Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1459861Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1459952Z layer_outputs = layer_module( 2025-09-07T07:43:06.1460294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1460395Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1460757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1460861Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1461222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1461328Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1461694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1461789Z value_states = self.v(current_states) 2025-09-07T07:43:06.1461793Z 2025-09-07T07:43:06.1461934Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1462220Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1462298Z return mod(**inputs) 2025-09-07T07:43:06.1462669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1462758Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1463126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1463210Z layer_outputs = layer_module( 2025-09-07T07:43:06.1463551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1463651Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1464016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1464121Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1464481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1464594Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1464955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1465092Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1465097Z 2025-09-07T07:43:06.1465235Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1465661Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1465744Z return mod(**inputs) 2025-09-07T07:43:06.1466110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1466390Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1466767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1466910Z layer_outputs = layer_module( 2025-09-07T07:43:06.1467256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1467353Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1467719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1467821Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1468243Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1468352Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1468716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1468856Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1468863Z 2025-09-07T07:43:06.1468996Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1469282Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1469360Z return mod(**inputs) 2025-09-07T07:43:06.1469723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1469814Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1470178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1470266Z layer_outputs = layer_module( 2025-09-07T07:43:06.1470609Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1470707Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1471072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1471176Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1471539Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1471645Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1472006Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1472144Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1472152Z 2025-09-07T07:43:06.1472286Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1472574Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1472652Z return mod(**inputs) 2025-09-07T07:43:06.1473018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1473111Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1473476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1473565Z layer_outputs = layer_module( 2025-09-07T07:43:06.1473904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1474004Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1474368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1474470Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1474925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1475041Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1475405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1475580Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1475585Z 2025-09-07T07:43:06.1475719Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1476007Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1476083Z return mod(**inputs) 2025-09-07T07:43:06.1476453Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1476585Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1476948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1477041Z layer_outputs = layer_module( 2025-09-07T07:43:06.1477382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1477481Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1477843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1477948Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1478308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1478416Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1478781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1478871Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1478875Z 2025-09-07T07:43:06.1478973Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1479110Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1479395Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1479476Z return mod(**inputs) 2025-09-07T07:43:06.1479842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1479934Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1480298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1480382Z layer_outputs = layer_module( 2025-09-07T07:43:06.1480724Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1480823Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1481190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1481301Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1481663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.1481791Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1482154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1482249Z return self.weight * hidden_states 2025-09-07T07:43:06.1482253Z 2025-09-07T07:43:06.1482386Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1482673Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1482751Z return mod(**inputs) 2025-09-07T07:43:06.1483117Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1483303Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1483674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1483798Z layer_outputs = layer_module( 2025-09-07T07:43:06.1484138Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1484236Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1484598Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1484708Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1485129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1485283Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1485650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.1485748Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.1485755Z 2025-09-07T07:43:06.1485889Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1486177Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1486254Z return mod(**inputs) 2025-09-07T07:43:06.1486626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1486715Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1487079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1487168Z layer_outputs = layer_module( 2025-09-07T07:43:06.1487511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1487608Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1487973Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1488090Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1488451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1488603Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1488967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.1489067Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.1489072Z 2025-09-07T07:43:06.1489210Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1489495Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1489574Z return mod(**inputs) 2025-09-07T07:43:06.1489942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1490032Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1490399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1490484Z layer_outputs = layer_module( 2025-09-07T07:43:06.1490837Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1490934Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1491296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1491413Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1491868Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1492027Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1492436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.1492532Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.1492537Z 2025-09-07T07:43:06.1492635Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1492769Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1493058Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1493134Z return mod(**inputs) 2025-09-07T07:43:06.1493550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1493640Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1494008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1494094Z layer_outputs = layer_module( 2025-09-07T07:43:06.1494438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1494537Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1494900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1495000Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1495368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.1495505Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1495868Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1495963Z return self.weight * hidden_states 2025-09-07T07:43:06.1495968Z 2025-09-07T07:43:06.1496101Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1496388Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1496465Z return mod(**inputs) 2025-09-07T07:43:06.1496832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1496920Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1497287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1497373Z layer_outputs = layer_module( 2025-09-07T07:43:06.1497716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1497816Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1498178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1498282Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1498645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1498747Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1499111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1499203Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1499208Z 2025-09-07T07:43:06.1499347Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1499632Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1499708Z return mod(**inputs) 2025-09-07T07:43:06.1500168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1500263Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1500630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1500753Z layer_outputs = layer_module( 2025-09-07T07:43:06.1501097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1501194Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1501557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1501659Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1502059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1502165Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1502530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1502622Z key_states = self.k(current_states) 2025-09-07T07:43:06.1502629Z 2025-09-07T07:43:06.1502766Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1503052Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1503131Z return mod(**inputs) 2025-09-07T07:43:06.1503496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1503584Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1503952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1504039Z layer_outputs = layer_module( 2025-09-07T07:43:06.1504384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1504483Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1504846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1504947Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1505309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1505413Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1505773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1505943Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1505950Z 2025-09-07T07:43:06.1506084Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1506369Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1506451Z return mod(**inputs) 2025-09-07T07:43:06.1506816Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1506910Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1507276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1507363Z layer_outputs = layer_module( 2025-09-07T07:43:06.1507704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1507799Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1508164Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1508268Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1508727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1508834Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1509201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1509417Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1509423Z 2025-09-07T07:43:06.1509557Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1509846Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1509921Z return mod(**inputs) 2025-09-07T07:43:06.1510286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1510417Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1510780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1510871Z layer_outputs = layer_module( 2025-09-07T07:43:06.1511211Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1511311Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1511672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1511771Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1512137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1512241Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1512606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1512776Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1512781Z 2025-09-07T07:43:06.1512918Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1513207Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1513285Z return mod(**inputs) 2025-09-07T07:43:06.1513654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1513744Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1514112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1514197Z layer_outputs = layer_module( 2025-09-07T07:43:06.1514537Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1514641Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1515010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1515111Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1515472Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1515576Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1515939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1516149Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1516154Z 2025-09-07T07:43:06.1516290Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1516575Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1516656Z return mod(**inputs) 2025-09-07T07:43:06.1517021Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1517200Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1517571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1517708Z layer_outputs = layer_module( 2025-09-07T07:43:06.1518048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1518143Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1518505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1518610Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1519012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1519117Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1519483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1519577Z value_states = self.v(current_states) 2025-09-07T07:43:06.1519586Z 2025-09-07T07:43:06.1519720Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1520003Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1520081Z return mod(**inputs) 2025-09-07T07:43:06.1520449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1520541Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1520907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1520993Z layer_outputs = layer_module( 2025-09-07T07:43:06.1521333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1521432Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1521794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1521897Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1522257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1522362Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1522722Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1522863Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1522870Z 2025-09-07T07:43:06.1523005Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1523293Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1523373Z return mod(**inputs) 2025-09-07T07:43:06.1523737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1523831Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1524196Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1524286Z layer_outputs = layer_module( 2025-09-07T07:43:06.1524625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1524719Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1525085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1525187Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1525640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1525752Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1526114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1526287Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1526293Z 2025-09-07T07:43:06.1526425Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1526715Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1526790Z return mod(**inputs) 2025-09-07T07:43:06.1527161Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1527293Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1527661Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1527752Z layer_outputs = layer_module( 2025-09-07T07:43:06.1528091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1528192Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1528553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1528654Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1529018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1529122Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1529486Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1529627Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1529631Z 2025-09-07T07:43:06.1529771Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1530056Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1530135Z return mod(**inputs) 2025-09-07T07:43:06.1530501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1530591Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1530959Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1531044Z layer_outputs = layer_module( 2025-09-07T07:43:06.1531384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1531488Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1531850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1531955Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1532318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1532421Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1532785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1532924Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1532929Z 2025-09-07T07:43:06.1533065Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1533350Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1533431Z return mod(**inputs) 2025-09-07T07:43:06.1533798Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1533885Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1534351Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1534475Z layer_outputs = layer_module( 2025-09-07T07:43:06.1534821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1534919Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1535281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1535384Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1535745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1535897Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1536264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1536355Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1536363Z 2025-09-07T07:43:06.1536460Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1536596Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1536883Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1536958Z return mod(**inputs) 2025-09-07T07:43:06.1537325Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1537412Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1537778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1537869Z layer_outputs = layer_module( 2025-09-07T07:43:06.1538208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1538313Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1538674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1538779Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1539143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:43:06.1539280Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1539643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1539736Z return self.weight * hidden_states 2025-09-07T07:43:06.1539743Z 2025-09-07T07:43:06.1539879Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1540163Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1540242Z return mod(**inputs) 2025-09-07T07:43:06.1540610Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1540701Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1541069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1541153Z layer_outputs = layer_module( 2025-09-07T07:43:06.1541492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1541591Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1541953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1542059Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1542513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1542628Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1542995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1543123Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1543128Z 2025-09-07T07:43:06.1543263Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1543550Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1543628Z return mod(**inputs) 2025-09-07T07:43:06.1543992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1544123Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1544493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1544581Z layer_outputs = layer_module( 2025-09-07T07:43:06.1544923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1545021Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1545381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1545486Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1545847Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1545955Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1546317Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1546410Z key_states = self.k(current_states) 2025-09-07T07:43:06.1546418Z 2025-09-07T07:43:06.1546552Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1546841Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1546919Z return mod(**inputs) 2025-09-07T07:43:06.1547285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1547375Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1547740Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1547824Z layer_outputs = layer_module( 2025-09-07T07:43:06.1548171Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1548269Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1548634Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1548741Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1549102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1549215Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1549576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1549746Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1549751Z 2025-09-07T07:43:06.1549883Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1550169Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1550247Z return mod(**inputs) 2025-09-07T07:43:06.1550613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1550708Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1551171Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1551296Z layer_outputs = layer_module( 2025-09-07T07:43:06.1551637Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1551732Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1552097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1552199Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1552562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1552718Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1553079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1553256Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1553261Z 2025-09-07T07:43:06.1553394Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1553682Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1553759Z return mod(**inputs) 2025-09-07T07:43:06.1554127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1554217Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1554582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1554670Z layer_outputs = layer_module( 2025-09-07T07:43:06.1555011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1555111Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1555476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1555581Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1555946Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1556053Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1556419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1556587Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1556594Z 2025-09-07T07:43:06.1556731Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1557014Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1557089Z return mod(**inputs) 2025-09-07T07:43:06.1557464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1557554Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1557924Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1558010Z layer_outputs = layer_module( 2025-09-07T07:43:06.1558349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1558450Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1558812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1558919Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1559280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1559479Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1559849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1560094Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1560099Z 2025-09-07T07:43:06.1560236Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1560520Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1560598Z return mod(**inputs) 2025-09-07T07:43:06.1560965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1561086Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1561457Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1561546Z layer_outputs = layer_module( 2025-09-07T07:43:06.1561889Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1561986Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1562348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1562453Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1562815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1562924Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1563285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1563383Z value_states = self.v(current_states) 2025-09-07T07:43:06.1563388Z 2025-09-07T07:43:06.1563522Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1563808Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1563888Z return mod(**inputs) 2025-09-07T07:43:06.1564257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1564349Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1564715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1564800Z layer_outputs = layer_module( 2025-09-07T07:43:06.1565144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1565244Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1565704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1565811Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1566173Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1566286Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1566648Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1566788Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1566793Z 2025-09-07T07:43:06.1566926Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1567213Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1567290Z return mod(**inputs) 2025-09-07T07:43:06.1567655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1567747Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1568277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1568413Z layer_outputs = layer_module( 2025-09-07T07:43:06.1568809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1568906Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1569270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1569373Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1569737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1569909Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1570272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1570413Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1570418Z 2025-09-07T07:43:06.1570555Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1570845Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1570921Z return mod(**inputs) 2025-09-07T07:43:06.1571289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1571377Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1571742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1571839Z layer_outputs = layer_module( 2025-09-07T07:43:06.1572180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1572281Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1572646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1572751Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1573116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1573223Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1573586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1573723Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1573727Z 2025-09-07T07:43:06.1573867Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1574150Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1574226Z return mod(**inputs) 2025-09-07T07:43:06.1574602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1574692Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1575064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1575151Z layer_outputs = layer_module( 2025-09-07T07:43:06.1575492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1575591Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1575955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1576063Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1576423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1576620Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1576989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1577161Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1577166Z 2025-09-07T07:43:06.1577303Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1577587Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1577666Z return mod(**inputs) 2025-09-07T07:43:06.1578032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1578161Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1578528Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1578613Z layer_outputs = layer_module( 2025-09-07T07:43:06.1578960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1579055Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1579421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1579526Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1579886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1579995Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1580355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1580452Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1580456Z 2025-09-07T07:43:06.1580590Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1580879Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1580957Z return mod(**inputs) 2025-09-07T07:43:06.1581323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1581417Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1581781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1581865Z layer_outputs = layer_module( 2025-09-07T07:43:06.1582210Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1582305Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1582670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1582772Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1583135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 647, in forward 2025-09-07T07:43:06.1583313Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-09-07T07:43:06.1583320Z 2025-09-07T07:43:06.1583419Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1583555Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1583841Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1583920Z return mod(**inputs) 2025-09-07T07:43:06.1584291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1584384Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1584751Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1584836Z layer_outputs = layer_module( 2025-09-07T07:43:06.1585314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1585450Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1585817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1585931Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1586295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.1586424Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1586787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1586933Z return self.weight * hidden_states 2025-09-07T07:43:06.1586938Z 2025-09-07T07:43:06.1587073Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1587360Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1587438Z return mod(**inputs) 2025-09-07T07:43:06.1587805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1587896Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1588261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1588346Z layer_outputs = layer_module( 2025-09-07T07:43:06.1588689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1588786Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1589153Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1589268Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1589630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1589788Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1590151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.1590251Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.1590256Z 2025-09-07T07:43:06.1590389Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1590678Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1590758Z return mod(**inputs) 2025-09-07T07:43:06.1591122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1591215Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1591584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1591673Z layer_outputs = layer_module( 2025-09-07T07:43:06.1592018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1592113Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1592481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1592592Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1592956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1593113Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1593477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.1593674Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.1593680Z 2025-09-07T07:43:06.1593819Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1594142Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1594218Z return mod(**inputs) 2025-09-07T07:43:06.1594585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1594674Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1595038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1595162Z layer_outputs = layer_module( 2025-09-07T07:43:06.1595501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1595600Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1595968Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1596082Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1596444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1596595Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1596959Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.1597057Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.1597062Z 2025-09-07T07:43:06.1597162Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1597297Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1597582Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1597663Z return mod(**inputs) 2025-09-07T07:43:06.1598029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1598123Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1598487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1598573Z layer_outputs = layer_module( 2025-09-07T07:43:06.1598915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1599014Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1599378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1599480Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1599844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.1599981Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1600341Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1600442Z return self.weight * hidden_states 2025-09-07T07:43:06.1600447Z 2025-09-07T07:43:06.1600581Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1600870Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1600945Z return mod(**inputs) 2025-09-07T07:43:06.1601312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1601405Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1601770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1602000Z layer_outputs = layer_module( 2025-09-07T07:43:06.1602354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1602491Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1602856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1602957Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1603323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1603427Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1603794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1603933Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1603938Z 2025-09-07T07:43:06.1604073Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1604367Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1604444Z return mod(**inputs) 2025-09-07T07:43:06.1604817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1604905Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1605270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1605356Z layer_outputs = layer_module( 2025-09-07T07:43:06.1605697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1605797Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1606160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1606262Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1606631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1606736Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1607100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1607194Z key_states = self.k(current_states) 2025-09-07T07:43:06.1607199Z 2025-09-07T07:43:06.1607337Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1607623Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1607702Z return mod(**inputs) 2025-09-07T07:43:06.1608074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1608165Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1608533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1608620Z layer_outputs = layer_module( 2025-09-07T07:43:06.1608962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1609060Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1609420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1609524Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1609887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1609996Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1610361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1610626Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1610631Z 2025-09-07T07:43:06.1610779Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1611100Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1611180Z return mod(**inputs) 2025-09-07T07:43:06.1611547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1611636Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1612007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1612151Z layer_outputs = layer_module( 2025-09-07T07:43:06.1612494Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1612591Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1612961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1613066Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1613428Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1613536Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1613897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1614068Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1614072Z 2025-09-07T07:43:06.1614208Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1614493Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1614573Z return mod(**inputs) 2025-09-07T07:43:06.1614942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1615034Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1615403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1615489Z layer_outputs = layer_module( 2025-09-07T07:43:06.1615834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1615930Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1616297Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1616401Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1616765Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1616873Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1617234Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1617406Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1617411Z 2025-09-07T07:43:06.1617545Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1617833Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1617909Z return mod(**inputs) 2025-09-07T07:43:06.1618276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1618371Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1618736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1618824Z layer_outputs = layer_module( 2025-09-07T07:43:06.1619261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1619365Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1619765Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1619865Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1620232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1620335Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1620701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1620954Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1620959Z 2025-09-07T07:43:06.1621092Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1621383Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1621461Z return mod(**inputs) 2025-09-07T07:43:06.1621831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1621920Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1622286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1622374Z layer_outputs = layer_module( 2025-09-07T07:43:06.1622711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1622813Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1623174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1623280Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1623640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1623744Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1624110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1624205Z value_states = self.v(current_states) 2025-09-07T07:43:06.1624209Z 2025-09-07T07:43:06.1624346Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1624631Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1624707Z return mod(**inputs) 2025-09-07T07:43:06.1625078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1625166Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1625538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1625623Z layer_outputs = layer_module( 2025-09-07T07:43:06.1625966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1626064Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1626429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1626531Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1626891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1627000Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1627360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1627593Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1627599Z 2025-09-07T07:43:06.1627742Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1628074Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1628152Z return mod(**inputs) 2025-09-07T07:43:06.1628521Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1628608Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1628976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1629060Z layer_outputs = layer_module( 2025-09-07T07:43:06.1629441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1629539Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1629908Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1630009Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1630373Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1630480Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1630846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1630986Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1630992Z 2025-09-07T07:43:06.1631124Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1631410Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1631490Z return mod(**inputs) 2025-09-07T07:43:06.1631856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1631949Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1632315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1632402Z layer_outputs = layer_module( 2025-09-07T07:43:06.1632747Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1632843Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1633210Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1633314Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1633680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1633784Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1634150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1634294Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1634299Z 2025-09-07T07:43:06.1634431Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1634721Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1634798Z return mod(**inputs) 2025-09-07T07:43:06.1635166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1635259Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1635629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1635719Z layer_outputs = layer_module( 2025-09-07T07:43:06.1636159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1636270Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1636669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1636769Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1637134Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1637237Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1637605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1637786Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1637791Z 2025-09-07T07:43:06.1637925Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1638217Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1638293Z return mod(**inputs) 2025-09-07T07:43:06.1638663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1638754Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1639118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1639206Z layer_outputs = layer_module( 2025-09-07T07:43:06.1639546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1639649Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1640015Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1640118Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1640482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1640585Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1640953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1641045Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1641050Z 2025-09-07T07:43:06.1641150Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1641284Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1641569Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1641649Z return mod(**inputs) 2025-09-07T07:43:06.1642018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1642109Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1642476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1642562Z layer_outputs = layer_module( 2025-09-07T07:43:06.1642907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1643002Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1643366Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1643469Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1643834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:43:06.1643972Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1644334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1644524Z return self.weight * hidden_states 2025-09-07T07:43:06.1644530Z 2025-09-07T07:43:06.1644669Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1645004Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1645082Z return mod(**inputs) 2025-09-07T07:43:06.1645447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1645540Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1645905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1645993Z layer_outputs = layer_module( 2025-09-07T07:43:06.1646356Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1646454Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1646822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1646925Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1647292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1647400Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1647764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1647857Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1647861Z 2025-09-07T07:43:06.1647997Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1648289Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1648364Z return mod(**inputs) 2025-09-07T07:43:06.1648735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1648824Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1649190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1649284Z layer_outputs = layer_module( 2025-09-07T07:43:06.1649623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1649723Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1650084Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1650191Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1650555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1650662Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1651029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1651122Z key_states = self.k(current_states) 2025-09-07T07:43:06.1651128Z 2025-09-07T07:43:06.1651265Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1651549Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1651625Z return mod(**inputs) 2025-09-07T07:43:06.1651995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1652085Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1652463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1658315Z layer_outputs = layer_module( 2025-09-07T07:43:06.1658963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1659081Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1659454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1659603Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1659969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1660079Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1660443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1660657Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1660662Z 2025-09-07T07:43:06.1660801Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1661092Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1661173Z return mod(**inputs) 2025-09-07T07:43:06.1661540Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1661638Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1662014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1662102Z layer_outputs = layer_module( 2025-09-07T07:43:06.1662442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1662538Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1662907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1663009Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1663378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1663485Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1663851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1664020Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1664025Z 2025-09-07T07:43:06.1664159Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1664448Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1664524Z return mod(**inputs) 2025-09-07T07:43:06.1664893Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1664984Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1665433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1665523Z layer_outputs = layer_module( 2025-09-07T07:43:06.1665865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1665967Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1666333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1666436Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1666800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1666908Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1667274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1667442Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1667634Z 2025-09-07T07:43:06.1667780Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1668065Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1668198Z return mod(**inputs) 2025-09-07T07:43:06.1668572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1668660Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1669027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1669111Z layer_outputs = layer_module( 2025-09-07T07:43:06.1669518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1669620Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1669989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1670096Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1670458Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1670567Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1670928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1671138Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1671144Z 2025-09-07T07:43:06.1671284Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1671570Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1671648Z return mod(**inputs) 2025-09-07T07:43:06.1672018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1672107Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1672477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1672564Z layer_outputs = layer_module( 2025-09-07T07:43:06.1672908Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1673004Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1673368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1673472Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1673835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1673946Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1674312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1674410Z value_states = self.v(current_states) 2025-09-07T07:43:06.1674417Z 2025-09-07T07:43:06.1674551Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1674835Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1674913Z return mod(**inputs) 2025-09-07T07:43:06.1675281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1675376Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1675742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1675831Z layer_outputs = layer_module( 2025-09-07T07:43:06.1676266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1676373Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1676742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1676879Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1677244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1677352Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1677716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1677895Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1677901Z 2025-09-07T07:43:06.1678036Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1678327Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1678406Z return mod(**inputs) 2025-09-07T07:43:06.1678772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1678867Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1679233Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1679321Z layer_outputs = layer_module( 2025-09-07T07:43:06.1679661Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1679758Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1680123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1680227Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1680596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1680705Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1681073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1681209Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1681214Z 2025-09-07T07:43:06.1681346Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1681636Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1681711Z return mod(**inputs) 2025-09-07T07:43:06.1682077Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1682169Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1682538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1682627Z layer_outputs = layer_module( 2025-09-07T07:43:06.1682964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1683065Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1683423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1683527Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1683888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1683994Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1684361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1684499Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1684503Z 2025-09-07T07:43:06.1684731Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1685024Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1685133Z return mod(**inputs) 2025-09-07T07:43:06.1685504Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1685594Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1685963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1686048Z layer_outputs = layer_module( 2025-09-07T07:43:06.1686392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1686527Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1686895Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1687001Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1687362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1687472Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1687833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1687972Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1687977Z 2025-09-07T07:43:06.1688112Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1688395Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1688476Z return mod(**inputs) 2025-09-07T07:43:06.1688843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1688935Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1689302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1689388Z layer_outputs = layer_module( 2025-09-07T07:43:06.1689731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1689828Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1690191Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1690292Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1690657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1690766Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1691129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1691224Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1691231Z 2025-09-07T07:43:06.1691332Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1691467Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1691753Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1691830Z return mod(**inputs) 2025-09-07T07:43:06.1692198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1692289Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1692659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1692746Z layer_outputs = layer_module( 2025-09-07T07:43:06.1693170Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1693276Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1693639Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1693786Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1694149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.1694273Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1694638Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1694778Z return self.weight * hidden_states 2025-09-07T07:43:06.1694783Z 2025-09-07T07:43:06.1694920Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1695210Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1695287Z return mod(**inputs) 2025-09-07T07:43:06.1695657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1695749Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1696117Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1696203Z layer_outputs = layer_module( 2025-09-07T07:43:06.1696548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1696645Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1697010Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1697126Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1697491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1697646Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1698011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.1698110Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.1698115Z 2025-09-07T07:43:06.1698252Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1698536Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1698616Z return mod(**inputs) 2025-09-07T07:43:06.1698981Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1699073Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1699448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1699534Z layer_outputs = layer_module( 2025-09-07T07:43:06.1699878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1699976Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1700340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1700450Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1700814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1700969Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1701341Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.1701444Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.1701561Z 2025-09-07T07:43:06.1701706Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1702002Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1702125Z return mod(**inputs) 2025-09-07T07:43:06.1702501Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1702597Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1702969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1703059Z layer_outputs = layer_module( 2025-09-07T07:43:06.1703449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1703547Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1703928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1704042Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1704423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1704582Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1704961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.1705067Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.1705071Z 2025-09-07T07:43:06.1705210Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1705510Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1705588Z return mod(**inputs) 2025-09-07T07:43:06.1705972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1706065Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1706445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1706543Z layer_outputs = layer_module( 2025-09-07T07:43:06.1706895Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1706997Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1707372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1707484Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1707854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 343, in forward 2025-09-07T07:43:06.1708021Z hidden_states = hidden_states + self.dropout(forwarded_states) 2025-09-07T07:43:06.1708026Z 2025-09-07T07:43:06.1708132Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1708267Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1708555Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1708636Z return mod(**inputs) 2025-09-07T07:43:06.1709002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1709093Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1709461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1709545Z layer_outputs = layer_module( 2025-09-07T07:43:06.1709894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1709992Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1710443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1710550Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1710950Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.1711088Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1711454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1711553Z return self.weight * hidden_states 2025-09-07T07:43:06.1711559Z 2025-09-07T07:43:06.1711693Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1712013Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1712089Z return mod(**inputs) 2025-09-07T07:43:06.1712461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1712555Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1712923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1713014Z layer_outputs = layer_module( 2025-09-07T07:43:06.1713358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1713461Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1713825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1713929Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1714295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1714400Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1714772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1714867Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1714872Z 2025-09-07T07:43:06.1715004Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1715300Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1715376Z return mod(**inputs) 2025-09-07T07:43:06.1715750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1715840Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1716208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1716296Z layer_outputs = layer_module( 2025-09-07T07:43:06.1716642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1716744Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1717110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1717218Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1717582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1717685Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1718049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1718143Z key_states = self.k(current_states) 2025-09-07T07:43:06.1718148Z 2025-09-07T07:43:06.1718283Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1718575Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1718753Z return mod(**inputs) 2025-09-07T07:43:06.1719129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1719256Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1719629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1719713Z layer_outputs = layer_module( 2025-09-07T07:43:06.1720059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1720156Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1720564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1720670Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1721036Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1721142Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1721505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1721676Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1721681Z 2025-09-07T07:43:06.1721819Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1722103Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1722184Z return mod(**inputs) 2025-09-07T07:43:06.1722554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1722644Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1723018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1723104Z layer_outputs = layer_module( 2025-09-07T07:43:06.1723450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1723549Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1723913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1724015Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1724375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1724482Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1724846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1725020Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1725025Z 2025-09-07T07:43:06.1725164Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1725449Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1725530Z return mod(**inputs) 2025-09-07T07:43:06.1725894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1725985Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1726350Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1726436Z layer_outputs = layer_module( 2025-09-07T07:43:06.1726777Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1726875Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1727333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1727443Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1727811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1727960Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1728322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1728496Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1728500Z 2025-09-07T07:43:06.1728636Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1728925Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1729034Z return mod(**inputs) 2025-09-07T07:43:06.1729400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1729497Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1729862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1729951Z layer_outputs = layer_module( 2025-09-07T07:43:06.1730292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1730391Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1730753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1730852Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1731220Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1731323Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1731700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1731914Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1731921Z 2025-09-07T07:43:06.1732056Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1732348Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1732424Z return mod(**inputs) 2025-09-07T07:43:06.1732801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1732891Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1733264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1733351Z layer_outputs = layer_module( 2025-09-07T07:43:06.1733694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1733794Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1734157Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1734263Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1734632Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1734739Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1735107Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1735213Z value_states = self.v(current_states) 2025-09-07T07:43:06.1735219Z 2025-09-07T07:43:06.1735361Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1735648Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1735827Z return mod(**inputs) 2025-09-07T07:43:06.1736204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1736329Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1736711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1736795Z layer_outputs = layer_module( 2025-09-07T07:43:06.1737140Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1737240Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1737603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1737746Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1738110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1738219Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1738582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1738721Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1738730Z 2025-09-07T07:43:06.1738863Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1739148Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1739229Z return mod(**inputs) 2025-09-07T07:43:06.1739606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1739700Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1740065Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1740162Z layer_outputs = layer_module( 2025-09-07T07:43:06.1740505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1740608Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1740974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1741077Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1741442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1741545Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1741915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1742055Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1742059Z 2025-09-07T07:43:06.1742204Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1742493Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1742576Z return mod(**inputs) 2025-09-07T07:43:06.1742947Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1743037Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1743412Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1743499Z layer_outputs = layer_module( 2025-09-07T07:43:06.1743843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1743945Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1744311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1744516Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1744889Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1745029Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1745390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1745527Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1745537Z 2025-09-07T07:43:06.1745672Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1745956Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1746069Z return mod(**inputs) 2025-09-07T07:43:06.1746434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1746529Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1746896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1746983Z layer_outputs = layer_module( 2025-09-07T07:43:06.1747327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1747423Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1747789Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1747890Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1748249Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1748358Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1748723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1748866Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1748870Z 2025-09-07T07:43:06.1749008Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1749295Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1749372Z return mod(**inputs) 2025-09-07T07:43:06.1749738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1749830Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1750197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1750289Z layer_outputs = layer_module( 2025-09-07T07:43:06.1750630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1750731Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1751100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1751204Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1751570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1751675Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1752047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1752142Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1752150Z 2025-09-07T07:43:06.1752248Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1752387Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1752673Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1752865Z return mod(**inputs) 2025-09-07T07:43:06.1753239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1753365Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1753735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1753822Z layer_outputs = layer_module( 2025-09-07T07:43:06.1754166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1754263Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1754627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1754775Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1755142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:43:06.1755282Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1755645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1755738Z return self.weight * hidden_states 2025-09-07T07:43:06.1755746Z 2025-09-07T07:43:06.1755878Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1756162Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1756242Z return mod(**inputs) 2025-09-07T07:43:06.1756607Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1756701Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1757067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1757154Z layer_outputs = layer_module( 2025-09-07T07:43:06.1757498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1757599Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1757964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1758067Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1758428Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1758540Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1758905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1759004Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1759008Z 2025-09-07T07:43:06.1759145Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1759433Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1759512Z return mod(**inputs) 2025-09-07T07:43:06.1759878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1759972Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1760341Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1760432Z layer_outputs = layer_module( 2025-09-07T07:43:06.1760774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1760874Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1761239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1761452Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1761824Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1761967Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1762328Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1762425Z key_states = self.k(current_states) 2025-09-07T07:43:06.1762430Z 2025-09-07T07:43:06.1762566Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1762856Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1762966Z return mod(**inputs) 2025-09-07T07:43:06.1763336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1763426Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1763797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1763890Z layer_outputs = layer_module( 2025-09-07T07:43:06.1764232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1764331Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1764694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1764795Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1765158Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1765369Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1765740Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1765915Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1765919Z 2025-09-07T07:43:06.1766055Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1766342Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1766418Z return mod(**inputs) 2025-09-07T07:43:06.1766789Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1766879Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1767249Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1767338Z layer_outputs = layer_module( 2025-09-07T07:43:06.1767677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1767785Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1768149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1768257Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1768618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1768725Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1769092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1769262Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1769269Z 2025-09-07T07:43:06.1769407Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1769690Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1769769Z return mod(**inputs) 2025-09-07T07:43:06.1770323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1770489Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1770860Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1770945Z layer_outputs = layer_module( 2025-09-07T07:43:06.1771290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1771387Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1771751Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1771935Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1772300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1772417Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1772780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1772955Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1772960Z 2025-09-07T07:43:06.1773096Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1773381Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1773461Z return mod(**inputs) 2025-09-07T07:43:06.1773826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1773920Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1774285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1774371Z layer_outputs = layer_module( 2025-09-07T07:43:06.1774719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1774818Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1775185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1775290Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1775654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1775764Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1776130Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1776346Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1776351Z 2025-09-07T07:43:06.1776492Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1776782Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1776862Z return mod(**inputs) 2025-09-07T07:43:06.1777228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1777321Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1777687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1777775Z layer_outputs = layer_module( 2025-09-07T07:43:06.1778115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1778216Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1778584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1778782Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1779156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1779299Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1779664Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1779759Z value_states = self.v(current_states) 2025-09-07T07:43:06.1779764Z 2025-09-07T07:43:06.1779900Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1780189Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1780305Z return mod(**inputs) 2025-09-07T07:43:06.1780680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1780769Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1781143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1781233Z layer_outputs = layer_module( 2025-09-07T07:43:06.1781578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1781680Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1782043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1782146Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1782511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1782621Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1782989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1783128Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1783133Z 2025-09-07T07:43:06.1783274Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1783561Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1783637Z return mod(**inputs) 2025-09-07T07:43:06.1784011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1784100Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1784470Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1784558Z layer_outputs = layer_module( 2025-09-07T07:43:06.1784900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1784999Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1785368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1785476Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1785837Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1785946Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1786308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1786446Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1786450Z 2025-09-07T07:43:06.1786590Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1786876Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1786954Z return mod(**inputs) 2025-09-07T07:43:06.1787419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1787517Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1787930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1788015Z layer_outputs = layer_module( 2025-09-07T07:43:06.1788361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1788458Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1788820Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1789009Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1789371Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1789483Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1789844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1789987Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1789992Z 2025-09-07T07:43:06.1790126Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1790413Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1790494Z return mod(**inputs) 2025-09-07T07:43:06.1790859Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1790952Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1791320Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1791406Z layer_outputs = layer_module( 2025-09-07T07:43:06.1791752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1791848Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1792216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1792320Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1792684Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1792792Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1793155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1793300Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1793305Z 2025-09-07T07:43:06.1793440Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1793730Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1793806Z return mod(**inputs) 2025-09-07T07:43:06.1794174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1794266Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1794633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1794720Z layer_outputs = layer_module( 2025-09-07T07:43:06.1795059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1795156Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1795524Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1795625Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1796112Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1796260Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1796627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1796716Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1796721Z 2025-09-07T07:43:06.1796816Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1796955Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1797242Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1797365Z return mod(**inputs) 2025-09-07T07:43:06.1797730Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1797819Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1798190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1798277Z layer_outputs = layer_module( 2025-09-07T07:43:06.1798619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1798717Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1799080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1799196Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1799558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.1799689Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1800053Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1800147Z return self.weight * hidden_states 2025-09-07T07:43:06.1800152Z 2025-09-07T07:43:06.1800287Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1800575Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1800653Z return mod(**inputs) 2025-09-07T07:43:06.1801021Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1801112Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1801478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1801565Z layer_outputs = layer_module( 2025-09-07T07:43:06.1801910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1802010Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1802377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1802491Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1802858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1803012Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1803376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.1803479Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.1803485Z 2025-09-07T07:43:06.1803620Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1803910Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1803985Z return mod(**inputs) 2025-09-07T07:43:06.1804453Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1804588Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1804955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1805046Z layer_outputs = layer_module( 2025-09-07T07:43:06.1805387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1805484Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1805849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1806005Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1806368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1806525Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1806892Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.1806994Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.1806998Z 2025-09-07T07:43:06.1807131Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1807419Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1807495Z return mod(**inputs) 2025-09-07T07:43:06.1807863Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1807954Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1808321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1808411Z layer_outputs = layer_module( 2025-09-07T07:43:06.1808757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1808860Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1809223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1809337Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1809698Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1809848Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1810215Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.1810312Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.1810317Z 2025-09-07T07:43:06.1810418Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1810554Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1810843Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1810924Z return mod(**inputs) 2025-09-07T07:43:06.1811288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1811381Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1811752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1811837Z layer_outputs = layer_module( 2025-09-07T07:43:06.1812178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1812276Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1812739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1812848Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1813215Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.1813388Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1813750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1813845Z return self.weight * hidden_states 2025-09-07T07:43:06.1813850Z 2025-09-07T07:43:06.1813984Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1814275Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1814391Z return mod(**inputs) 2025-09-07T07:43:06.1814758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1814855Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1815221Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1815310Z layer_outputs = layer_module( 2025-09-07T07:43:06.1815647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1815744Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1816109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1816209Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1816573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1816678Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1817044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1817137Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1817141Z 2025-09-07T07:43:06.1817277Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1817565Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1817641Z return mod(**inputs) 2025-09-07T07:43:06.1818011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1818102Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1818469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1818562Z layer_outputs = layer_module( 2025-09-07T07:43:06.1818902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1819007Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1819370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1819475Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1819836Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1819938Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1820303Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1820393Z key_states = self.k(current_states) 2025-09-07T07:43:06.1820401Z 2025-09-07T07:43:06.1820537Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1820822Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1820899Z return mod(**inputs) 2025-09-07T07:43:06.1821369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1821470Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1821877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1821963Z layer_outputs = layer_module( 2025-09-07T07:43:06.1822308Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1822408Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1822771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1822927Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1823291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1823401Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1823764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1823939Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1823944Z 2025-09-07T07:43:06.1824081Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1824368Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1824450Z return mod(**inputs) 2025-09-07T07:43:06.1824816Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1824907Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1825275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1825360Z layer_outputs = layer_module( 2025-09-07T07:43:06.1825709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1825808Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1826176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1826276Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1826636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1826741Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1827107Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1827282Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1827287Z 2025-09-07T07:43:06.1827423Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1827711Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1827793Z return mod(**inputs) 2025-09-07T07:43:06.1828162Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1828253Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1828622Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1828706Z layer_outputs = layer_module( 2025-09-07T07:43:06.1829049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1829148Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1829515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1829713Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1830085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1830220Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1830586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1830764Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1830768Z 2025-09-07T07:43:06.1830903Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1831191Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1831300Z return mod(**inputs) 2025-09-07T07:43:06.1831669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1831762Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1832133Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1832226Z layer_outputs = layer_module( 2025-09-07T07:43:06.1832565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1832663Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1833026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1833126Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1833492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1833601Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1833969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1834181Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1834186Z 2025-09-07T07:43:06.1834324Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1834614Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1834690Z return mod(**inputs) 2025-09-07T07:43:06.1835061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1835151Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1835520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1835611Z layer_outputs = layer_module( 2025-09-07T07:43:06.1835952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1836056Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1836419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1836526Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1836885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1836988Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1837354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1837449Z value_states = self.v(current_states) 2025-09-07T07:43:06.1837454Z 2025-09-07T07:43:06.1837595Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1837877Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1837954Z return mod(**inputs) 2025-09-07T07:43:06.1838420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1838516Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1838926Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1839011Z layer_outputs = layer_module( 2025-09-07T07:43:06.1839356Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1839455Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1839818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1839964Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1840330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1840442Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1840804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1840945Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1840951Z 2025-09-07T07:43:06.1841088Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1841371Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1841449Z return mod(**inputs) 2025-09-07T07:43:06.1841814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1841905Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1842273Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1842359Z layer_outputs = layer_module( 2025-09-07T07:43:06.1842707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1842807Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1843173Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1843275Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1843636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1843743Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1844107Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1844249Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1844254Z 2025-09-07T07:43:06.1844388Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1844678Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1844758Z return mod(**inputs) 2025-09-07T07:43:06.1845127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1845221Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1845586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1845673Z layer_outputs = layer_module( 2025-09-07T07:43:06.1846015Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1846114Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1846479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1846580Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1847034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1847177Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1847541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1847682Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1847687Z 2025-09-07T07:43:06.1847821Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1848111Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1848188Z return mod(**inputs) 2025-09-07T07:43:06.1848598Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1848690Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1849059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1849149Z layer_outputs = layer_module( 2025-09-07T07:43:06.1849492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1849593Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1849959Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1850060Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1850428Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1850532Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1850897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1851039Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1851044Z 2025-09-07T07:43:06.1851177Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1851469Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1851546Z return mod(**inputs) 2025-09-07T07:43:06.1851918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1852007Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1852374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1852459Z layer_outputs = layer_module( 2025-09-07T07:43:06.1852799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1852899Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1853265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1853369Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1853731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1853834Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1854202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1854294Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1854299Z 2025-09-07T07:43:06.1854436Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1854725Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1854801Z return mod(**inputs) 2025-09-07T07:43:06.1855266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1855363Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1855736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1855886Z layer_outputs = layer_module( 2025-09-07T07:43:06.1856230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1856329Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1856691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1856795Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1857198Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T07:43:06.1857376Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T07:43:06.1857384Z 2025-09-07T07:43:06.1857480Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1857612Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1857902Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1857978Z return mod(**inputs) 2025-09-07T07:43:06.1858350Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1858440Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1858805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1858895Z layer_outputs = layer_module( 2025-09-07T07:43:06.1859237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1859339Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1859705Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1859813Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1860171Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:43:06.1860307Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1860670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1860764Z return self.weight * hidden_states 2025-09-07T07:43:06.1860768Z 2025-09-07T07:43:06.1860905Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1861193Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1861269Z return mod(**inputs) 2025-09-07T07:43:06.1861639Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1861728Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1862099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1862183Z layer_outputs = layer_module( 2025-09-07T07:43:06.1862526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1862623Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1862988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1863100Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1863464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1863577Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1864034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1864169Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1864175Z 2025-09-07T07:43:06.1864316Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1864601Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1864681Z return mod(**inputs) 2025-09-07T07:43:06.1865052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1865140Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1865681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1865767Z layer_outputs = layer_module( 2025-09-07T07:43:06.1866116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1866214Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1866582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1866685Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1867048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1867160Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1867522Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1867619Z key_states = self.k(current_states) 2025-09-07T07:43:06.1867624Z 2025-09-07T07:43:06.1867759Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1868048Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1868126Z return mod(**inputs) 2025-09-07T07:43:06.1868493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1868587Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1868953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1869040Z layer_outputs = layer_module( 2025-09-07T07:43:06.1869382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1869478Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1869847Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1869947Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1870314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1870423Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1870784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1870957Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1870961Z 2025-09-07T07:43:06.1871097Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1871388Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1871464Z return mod(**inputs) 2025-09-07T07:43:06.1871834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1871930Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1872478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1872586Z layer_outputs = layer_module( 2025-09-07T07:43:06.1872926Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1873080Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1873443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1873545Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1873913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1874076Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1874441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1874610Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1874620Z 2025-09-07T07:43:06.1874755Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1875045Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1875127Z return mod(**inputs) 2025-09-07T07:43:06.1875500Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1875589Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1875958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1876043Z layer_outputs = layer_module( 2025-09-07T07:43:06.1876384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1876483Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1876847Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1876953Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1877315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1877421Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1877785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1877954Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1877958Z 2025-09-07T07:43:06.1878095Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1878382Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1878457Z return mod(**inputs) 2025-09-07T07:43:06.1878826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1878916Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1879283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1879370Z layer_outputs = layer_module( 2025-09-07T07:43:06.1879713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1879810Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1880171Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1880280Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1880644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1880757Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1881208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1881428Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1881465Z 2025-09-07T07:43:06.1881609Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1881895Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1881976Z return mod(**inputs) 2025-09-07T07:43:06.1882352Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1882445Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1882869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1882956Z layer_outputs = layer_module( 2025-09-07T07:43:06.1883312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1883414Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1883795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1883903Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1884273Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1884390Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1884765Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1884868Z value_states = self.v(current_states) 2025-09-07T07:43:06.1884874Z 2025-09-07T07:43:06.1885014Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1885311Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1885397Z return mod(**inputs) 2025-09-07T07:43:06.1885779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1885878Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1886256Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1886347Z layer_outputs = layer_module( 2025-09-07T07:43:06.1886701Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1886802Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1887185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1887290Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1887677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1887789Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1888163Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1888307Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1888312Z 2025-09-07T07:43:06.1888448Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1888746Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1888823Z return mod(**inputs) 2025-09-07T07:43:06.1889204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1889298Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1889785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1889889Z layer_outputs = layer_module( 2025-09-07T07:43:06.1890245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1890384Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1890755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1890858Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1891226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1891334Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1891741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1891882Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1891890Z 2025-09-07T07:43:06.1892027Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1892319Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1892400Z return mod(**inputs) 2025-09-07T07:43:06.1892775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1892865Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1893238Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1893324Z layer_outputs = layer_module( 2025-09-07T07:43:06.1893672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1893771Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1894139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1894245Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1894605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1894714Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1895081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1895218Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1895223Z 2025-09-07T07:43:06.1895362Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1895652Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1895732Z return mod(**inputs) 2025-09-07T07:43:06.1896102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1896192Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1896562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1896650Z layer_outputs = layer_module( 2025-09-07T07:43:06.1896994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1897091Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1897454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1897561Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1897930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1898043Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1898525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1898678Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1898722Z 2025-09-07T07:43:06.1898856Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1899141Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1899221Z return mod(**inputs) 2025-09-07T07:43:06.1899592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1899687Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1900054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1900190Z layer_outputs = layer_module( 2025-09-07T07:43:06.1900541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1900639Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1901007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1901112Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1901476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1901588Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1901955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1902050Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1902057Z 2025-09-07T07:43:06.1902155Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1902291Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1902585Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1902662Z return mod(**inputs) 2025-09-07T07:43:06.1903036Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1903129Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1903511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1903596Z layer_outputs = layer_module( 2025-09-07T07:43:06.1903935Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1904035Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1904399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1904516Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1904880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.1905006Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1905378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1905473Z return self.weight * hidden_states 2025-09-07T07:43:06.1905478Z 2025-09-07T07:43:06.1905615Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1905904Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1905985Z return mod(**inputs) 2025-09-07T07:43:06.1906355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1906443Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1906907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1907001Z layer_outputs = layer_module( 2025-09-07T07:43:06.1907350Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1907514Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1907877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1907992Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1908354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1908545Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1908910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.1909008Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.1909020Z 2025-09-07T07:43:06.1909153Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1909439Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1909522Z return mod(**inputs) 2025-09-07T07:43:06.1909890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1909982Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1910351Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1910437Z layer_outputs = layer_module( 2025-09-07T07:43:06.1910786Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1910884Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1911255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1911369Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1911739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1911896Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1912260Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.1912363Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.1912368Z 2025-09-07T07:43:06.1912502Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1912793Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1912870Z return mod(**inputs) 2025-09-07T07:43:06.1913242Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1913336Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1913702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1913793Z layer_outputs = layer_module( 2025-09-07T07:43:06.1914143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1914240Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1914624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.1914741Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.1915129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.1915286Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.1915751Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.1915857Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.1915893Z 2025-09-07T07:43:06.1915992Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1916132Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1916416Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1916496Z return mod(**inputs) 2025-09-07T07:43:06.1916872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1916997Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1917367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1917453Z layer_outputs = layer_module( 2025-09-07T07:43:06.1917801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1917898Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1918263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1918367Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1918728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.1918867Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1919228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1919323Z return self.weight * hidden_states 2025-09-07T07:43:06.1919330Z 2025-09-07T07:43:06.1919464Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1919750Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1919830Z return mod(**inputs) 2025-09-07T07:43:06.1920211Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1920311Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1920687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1920780Z layer_outputs = layer_module( 2025-09-07T07:43:06.1921128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1921229Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1921599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1921700Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1922065Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1922176Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1922547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1922644Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1922649Z 2025-09-07T07:43:06.1922782Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1923071Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1923148Z return mod(**inputs) 2025-09-07T07:43:06.1923519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1923615Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1924095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1924191Z layer_outputs = layer_module( 2025-09-07T07:43:06.1924613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1924709Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1925078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1925184Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1925554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1925703Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1926069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1926164Z key_states = self.k(current_states) 2025-09-07T07:43:06.1926173Z 2025-09-07T07:43:06.1926309Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1926597Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1926677Z return mod(**inputs) 2025-09-07T07:43:06.1927047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1927135Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1927506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1927597Z layer_outputs = layer_module( 2025-09-07T07:43:06.1927941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1928041Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1928407Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1928508Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1928879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1928982Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1929350Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1929521Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1929526Z 2025-09-07T07:43:06.1929662Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1929948Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1930024Z return mod(**inputs) 2025-09-07T07:43:06.1930400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1930490Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1930858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1930946Z layer_outputs = layer_module( 2025-09-07T07:43:06.1931286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1931391Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1931754Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1931860Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1932225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1932329Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1932791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1932968Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1933005Z 2025-09-07T07:43:06.1933146Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1933431Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1933513Z return mod(**inputs) 2025-09-07T07:43:06.1933881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1933970Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1934367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1934453Z layer_outputs = layer_module( 2025-09-07T07:43:06.1934801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1934898Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1935260Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1935367Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1935729Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1935836Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1936199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1936375Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1936380Z 2025-09-07T07:43:06.1936515Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1936804Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1936884Z return mod(**inputs) 2025-09-07T07:43:06.1937250Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1937344Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1937709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1937794Z layer_outputs = layer_module( 2025-09-07T07:43:06.1938138Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1938235Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1938602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1938705Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1939069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1939175Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1939541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1939754Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1939759Z 2025-09-07T07:43:06.1939895Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1940180Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1940257Z return mod(**inputs) 2025-09-07T07:43:06.1940626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1940720Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1941177Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1941274Z layer_outputs = layer_module( 2025-09-07T07:43:06.1941614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1941748Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1942116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1942216Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1942586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1942727Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1943092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1943187Z value_states = self.v(current_states) 2025-09-07T07:43:06.1943196Z 2025-09-07T07:43:06.1943332Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1943622Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1943703Z return mod(**inputs) 2025-09-07T07:43:06.1944072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1944161Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1944526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1944616Z layer_outputs = layer_module( 2025-09-07T07:43:06.1944956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1945056Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1945426Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1945527Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1945894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1945998Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1946364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1946504Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1946510Z 2025-09-07T07:43:06.1946648Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1946936Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1947016Z return mod(**inputs) 2025-09-07T07:43:06.1947386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1947479Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1947850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1947937Z layer_outputs = layer_module( 2025-09-07T07:43:06.1948277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1948381Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1948743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1948849Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1949213Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1949323Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1949767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1949913Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1949948Z 2025-09-07T07:43:06.1950088Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1950373Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1950456Z return mod(**inputs) 2025-09-07T07:43:06.1950821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1950910Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1951279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1951416Z layer_outputs = layer_module( 2025-09-07T07:43:06.1951765Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1951864Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1952226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1952332Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1952693Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1952801Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1953162Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1953304Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1953311Z 2025-09-07T07:43:06.1953447Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1953731Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1953815Z return mod(**inputs) 2025-09-07T07:43:06.1954182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1954275Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1954640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1954724Z layer_outputs = layer_module( 2025-09-07T07:43:06.1955070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1955168Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1955536Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1955639Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1956004Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1956110Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1956474Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.1956619Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.1956624Z 2025-09-07T07:43:06.1956758Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1957045Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1957121Z return mod(**inputs) 2025-09-07T07:43:06.1957488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1957585Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1957953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1958131Z layer_outputs = layer_module( 2025-09-07T07:43:06.1958481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1958612Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1958977Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.1959077Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.1959439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.1959542Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.1959940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.1960031Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.1960036Z 2025-09-07T07:43:06.1960138Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.1960276Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1960563Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1960645Z return mod(**inputs) 2025-09-07T07:43:06.1961011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1961101Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1961471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1961561Z layer_outputs = layer_module( 2025-09-07T07:43:06.1961906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1962003Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1962369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1962477Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1962838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:43:06.1962981Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.1963343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.1963441Z return self.weight * hidden_states 2025-09-07T07:43:06.1963446Z 2025-09-07T07:43:06.1963580Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1963867Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1963949Z return mod(**inputs) 2025-09-07T07:43:06.1964316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1964415Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1964783Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1964870Z layer_outputs = layer_module( 2025-09-07T07:43:06.1965213Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1965394Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1965769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1965871Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1966233Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1966345Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1966872Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.1966978Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.1967033Z 2025-09-07T07:43:06.1967173Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1967463Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1967540Z return mod(**inputs) 2025-09-07T07:43:06.1967906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1968000Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1968365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1968514Z layer_outputs = layer_module( 2025-09-07T07:43:06.1968854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1968955Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1969323Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1969428Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1969795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1969903Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1970270Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.1970362Z key_states = self.k(current_states) 2025-09-07T07:43:06.1970369Z 2025-09-07T07:43:06.1970503Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1970792Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1970867Z return mod(**inputs) 2025-09-07T07:43:06.1971241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1971332Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1971698Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1971789Z layer_outputs = layer_module( 2025-09-07T07:43:06.1972131Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1972232Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1972595Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1972698Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1973063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1973174Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1973541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1973711Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1973716Z 2025-09-07T07:43:06.1973856Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1974140Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1974216Z return mod(**inputs) 2025-09-07T07:43:06.1974585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1974676Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1975044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1975221Z layer_outputs = layer_module( 2025-09-07T07:43:06.1975571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1975709Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1976070Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1976177Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1976539Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1976649Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1977012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1977214Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1977218Z 2025-09-07T07:43:06.1977363Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1977647Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1977730Z return mod(**inputs) 2025-09-07T07:43:06.1978097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1978187Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1978559Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1978648Z layer_outputs = layer_module( 2025-09-07T07:43:06.1978992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1979092Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1979453Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1979561Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1979922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1980035Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1980395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.1980569Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.1980573Z 2025-09-07T07:43:06.1980706Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1980991Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1981071Z return mod(**inputs) 2025-09-07T07:43:06.1981437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1981534Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1981900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1981988Z layer_outputs = layer_module( 2025-09-07T07:43:06.1982333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1982428Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1982797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1982900Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1983265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1983375Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1983813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.1984031Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.1984080Z 2025-09-07T07:43:06.1984217Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1984507Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1984582Z return mod(**inputs) 2025-09-07T07:43:06.1984952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1989378Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1989853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1990065Z layer_outputs = layer_module( 2025-09-07T07:43:06.1990413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1990515Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1990886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1990992Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1991357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1991463Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1991830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.1991926Z value_states = self.v(current_states) 2025-09-07T07:43:06.1991933Z 2025-09-07T07:43:06.1992068Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1992357Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1992438Z return mod(**inputs) 2025-09-07T07:43:06.1992808Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1992899Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1993268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1993353Z layer_outputs = layer_module( 2025-09-07T07:43:06.1993697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1993790Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1994154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1994260Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1994621Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1994734Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1995094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1995236Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1995241Z 2025-09-07T07:43:06.1995374Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1995659Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1995739Z return mod(**inputs) 2025-09-07T07:43:06.1996106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1996201Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.1996569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.1996747Z layer_outputs = layer_module( 2025-09-07T07:43:06.1997095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.1997227Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.1997593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.1997696Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.1998063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.1998176Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.1998542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.1998732Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.1998739Z 2025-09-07T07:43:06.1998878Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.1999166Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.1999245Z return mod(**inputs) 2025-09-07T07:43:06.1999612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.1999704Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2000072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2000161Z layer_outputs = layer_module( 2025-09-07T07:43:06.2000500Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2000599Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2000963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2001071Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2001438Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2001548Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2001916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2002055Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2002060Z 2025-09-07T07:43:06.2002197Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2002487Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2002566Z return mod(**inputs) 2025-09-07T07:43:06.2002938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2003029Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2003397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2003489Z layer_outputs = layer_module( 2025-09-07T07:43:06.2003832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2003933Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2004296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2004399Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2004766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2004873Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2005239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.2005469Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.2005474Z 2025-09-07T07:43:06.2005618Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2005935Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2006012Z return mod(**inputs) 2025-09-07T07:43:06.2006386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2006479Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2006851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2006974Z layer_outputs = layer_module( 2025-09-07T07:43:06.2007314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2007416Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2007779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2007887Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2008249Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2008359Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2008723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.2008815Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.2008820Z 2025-09-07T07:43:06.2008962Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2009249Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2009330Z return mod(**inputs) 2025-09-07T07:43:06.2009700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2009789Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2010163Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2010250Z layer_outputs = layer_module( 2025-09-07T07:43:06.2010596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2010692Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2011055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2011162Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2011526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 647, in forward 2025-09-07T07:43:06.2011705Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-09-07T07:43:06.2011710Z 2025-09-07T07:43:06.2011808Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.2011950Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2012234Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2012310Z return mod(**inputs) 2025-09-07T07:43:06.2012677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2012766Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2013135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2013224Z layer_outputs = layer_module( 2025-09-07T07:43:06.2013565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2013751Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2014121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2014267Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2014629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.2014755Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.2015121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.2015215Z return self.weight * hidden_states 2025-09-07T07:43:06.2015252Z 2025-09-07T07:43:06.2015392Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2015677Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2015757Z return mod(**inputs) 2025-09-07T07:43:06.2016127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2016217Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2016589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2016674Z layer_outputs = layer_module( 2025-09-07T07:43:06.2017019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2017115Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2017476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2017593Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2017953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.2018111Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.2018472Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.2018574Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.2018579Z 2025-09-07T07:43:06.2018714Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2018998Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2019075Z return mod(**inputs) 2025-09-07T07:43:06.2019442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2019536Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2019908Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2019997Z layer_outputs = layer_module( 2025-09-07T07:43:06.2020342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2020440Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2020803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2020914Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2021273Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.2021430Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.2021792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.2021898Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.2021904Z 2025-09-07T07:43:06.2022149Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2022446Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2022560Z return mod(**inputs) 2025-09-07T07:43:06.2022925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2023021Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2023390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2023479Z layer_outputs = layer_module( 2025-09-07T07:43:06.2023819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2023957Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2024324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2024438Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2024804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.2024958Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.2025324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.2025420Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.2025424Z 2025-09-07T07:43:06.2025523Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.2025662Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2025945Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2026027Z return mod(**inputs) 2025-09-07T07:43:06.2026393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2026486Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2026853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2026941Z layer_outputs = layer_module( 2025-09-07T07:43:06.2027282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2027378Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2027743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2027847Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2028209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.2028348Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.2028716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.2028814Z return self.weight * hidden_states 2025-09-07T07:43:06.2028821Z 2025-09-07T07:43:06.2028953Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2029236Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2029315Z return mod(**inputs) 2025-09-07T07:43:06.2029678Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2029771Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2030136Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2030224Z layer_outputs = layer_module( 2025-09-07T07:43:06.2030570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2030759Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2031131Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2031264Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2031631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2031733Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2032095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.2032192Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.2032232Z 2025-09-07T07:43:06.2032368Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2032655Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2032736Z return mod(**inputs) 2025-09-07T07:43:06.2033100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2033194Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2033558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2033649Z layer_outputs = layer_module( 2025-09-07T07:43:06.2033991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2034088Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2034454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2034555Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2034923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2035025Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2035386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.2035481Z key_states = self.k(current_states) 2025-09-07T07:43:06.2035486Z 2025-09-07T07:43:06.2035619Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2035909Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2035985Z return mod(**inputs) 2025-09-07T07:43:06.2036354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2036446Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2036811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2036900Z layer_outputs = layer_module( 2025-09-07T07:43:06.2037242Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2037344Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2037710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2037810Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2038174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2038274Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2038637Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2038807Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2038812Z 2025-09-07T07:43:06.2039039Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2039332Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2039441Z return mod(**inputs) 2025-09-07T07:43:06.2039815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2039907Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2040276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2040361Z layer_outputs = layer_module( 2025-09-07T07:43:06.2040704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2040846Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2041209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2041314Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2041674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2041780Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2042140Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2042309Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2042313Z 2025-09-07T07:43:06.2042453Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2042739Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2042821Z return mod(**inputs) 2025-09-07T07:43:06.2043189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2043278Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2043654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2043742Z layer_outputs = layer_module( 2025-09-07T07:43:06.2044090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2044185Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2044546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2044651Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2045014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2045123Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2045484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2045659Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2045664Z 2025-09-07T07:43:06.2045801Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2046085Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2046165Z return mod(**inputs) 2025-09-07T07:43:06.2046531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2046626Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2046991Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2047079Z layer_outputs = layer_module( 2025-09-07T07:43:06.2047424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2047608Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2047981Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2048119Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2048483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2048586Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2048947Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.2049161Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.2049207Z 2025-09-07T07:43:06.2049343Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2049633Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2049707Z return mod(**inputs) 2025-09-07T07:43:06.2050080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2050174Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2050541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2050630Z layer_outputs = layer_module( 2025-09-07T07:43:06.2050970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2051065Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2051428Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2051532Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2051896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2052001Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2052366Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.2052461Z value_states = self.v(current_states) 2025-09-07T07:43:06.2052465Z 2025-09-07T07:43:06.2052598Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2052886Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2052963Z return mod(**inputs) 2025-09-07T07:43:06.2053333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2053425Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2053789Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2053878Z layer_outputs = layer_module( 2025-09-07T07:43:06.2054222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2054324Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2054685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2054788Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2055146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2055248Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2055615Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2055755Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2055760Z 2025-09-07T07:43:06.2055897Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2056297Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2056381Z return mod(**inputs) 2025-09-07T07:43:06.2056787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2056878Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2057248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2057333Z layer_outputs = layer_module( 2025-09-07T07:43:06.2057671Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2057809Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2058170Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2058279Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2058640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2058749Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2059109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2059245Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2059250Z 2025-09-07T07:43:06.2059386Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2059669Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2059749Z return mod(**inputs) 2025-09-07T07:43:06.2060113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2060201Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2060573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2060658Z layer_outputs = layer_module( 2025-09-07T07:43:06.2061004Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2061101Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2061467Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2061568Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2061927Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2062038Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2062398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2062542Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2062547Z 2025-09-07T07:43:06.2062681Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2062968Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2063047Z return mod(**inputs) 2025-09-07T07:43:06.2063416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2063509Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2063873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2063960Z layer_outputs = layer_module( 2025-09-07T07:43:06.2064305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2064401Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2064897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2065005Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2065484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2065586Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2065946Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.2066090Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.2066095Z 2025-09-07T07:43:06.2066228Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2066627Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2066704Z return mod(**inputs) 2025-09-07T07:43:06.2067075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2067169Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2067533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2067622Z layer_outputs = layer_module( 2025-09-07T07:43:06.2067960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2068059Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2068420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2068522Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2068888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2068993Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2069361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.2069455Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.2069460Z 2025-09-07T07:43:06.2069554Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.2069694Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2069979Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2070060Z return mod(**inputs) 2025-09-07T07:43:06.2070436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2070528Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2070896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2070981Z layer_outputs = layer_module( 2025-09-07T07:43:06.2071324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2071429Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2071791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2071898Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2072257Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:43:06.2072393Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.2072763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.2072858Z return self.weight * hidden_states 2025-09-07T07:43:06.2072863Z 2025-09-07T07:43:06.2073000Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2073419Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2073505Z return mod(**inputs) 2025-09-07T07:43:06.2073931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2074022Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2074392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2074481Z layer_outputs = layer_module( 2025-09-07T07:43:06.2074825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2074957Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2075321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2075430Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2075792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2075905Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2076268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.2076360Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.2076365Z 2025-09-07T07:43:06.2076503Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2076787Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2076867Z return mod(**inputs) 2025-09-07T07:43:06.2077236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2077329Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2077695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2077780Z layer_outputs = layer_module( 2025-09-07T07:43:06.2078123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2078220Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2078584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2078686Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2079048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2079161Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2079523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.2079620Z key_states = self.k(current_states) 2025-09-07T07:43:06.2079625Z 2025-09-07T07:43:06.2079759Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2080044Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2080122Z return mod(**inputs) 2025-09-07T07:43:06.2080487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2080577Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2080941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2081030Z layer_outputs = layer_module( 2025-09-07T07:43:06.2081375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2081470Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2081909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2082017Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2082414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2082522Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2082883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2083056Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2083060Z 2025-09-07T07:43:06.2083195Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2083531Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2083607Z return mod(**inputs) 2025-09-07T07:43:06.2083978Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2084069Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2084433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2084523Z layer_outputs = layer_module( 2025-09-07T07:43:06.2084862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2084961Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2085322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2085426Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2085791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2085898Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2086265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2086435Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2086440Z 2025-09-07T07:43:06.2086573Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2086860Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2086935Z return mod(**inputs) 2025-09-07T07:43:06.2087303Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2087392Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2087767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2087853Z layer_outputs = layer_module( 2025-09-07T07:43:06.2088195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2088295Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2088660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2088765Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2089126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2089234Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2089601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2089772Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2089777Z 2025-09-07T07:43:06.2089914Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2090290Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2090375Z return mod(**inputs) 2025-09-07T07:43:06.2090742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2090867Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2091236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2091322Z layer_outputs = layer_module( 2025-09-07T07:43:06.2091665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2091762Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2092158Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2092263Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2092630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2092740Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2093103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.2093315Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.2093324Z 2025-09-07T07:43:06.2093458Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2093743Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2093823Z return mod(**inputs) 2025-09-07T07:43:06.2094194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2094286Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2094653Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2094738Z layer_outputs = layer_module( 2025-09-07T07:43:06.2095084Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2095183Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2095550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2095653Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2096013Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2096127Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2096488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.2096588Z value_states = self.v(current_states) 2025-09-07T07:43:06.2096593Z 2025-09-07T07:43:06.2096726Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2097015Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2097091Z return mod(**inputs) 2025-09-07T07:43:06.2097458Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2097552Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2097916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2098004Z layer_outputs = layer_module( 2025-09-07T07:43:06.2098347Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2098444Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2098904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2099013Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2099413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2099520Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2099881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2100021Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2100026Z 2025-09-07T07:43:06.2100160Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2100490Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2100566Z return mod(**inputs) 2025-09-07T07:43:06.2100938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2101026Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2101390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2101481Z layer_outputs = layer_module( 2025-09-07T07:43:06.2101823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2101922Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2102283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2102388Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2102755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2102860Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2103229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2103367Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2103374Z 2025-09-07T07:43:06.2103511Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2103794Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2103869Z return mod(**inputs) 2025-09-07T07:43:06.2104240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2104328Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2104697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2104783Z layer_outputs = layer_module( 2025-09-07T07:43:06.2105126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2105226Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2105591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2105694Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2106055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2106161Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2106524Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2106663Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2106668Z 2025-09-07T07:43:06.2106804Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2107180Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2107267Z return mod(**inputs) 2025-09-07T07:43:06.2107633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2107759Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2108128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2108211Z layer_outputs = layer_module( 2025-09-07T07:43:06.2108555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2108650Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2109046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2109152Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2109517Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2109626Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2109989Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.2110129Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.2110136Z 2025-09-07T07:43:06.2110269Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2110555Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2110632Z return mod(**inputs) 2025-09-07T07:43:06.2110999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2111093Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2111461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2111547Z layer_outputs = layer_module( 2025-09-07T07:43:06.2111893Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2111994Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2112358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2112461Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2112819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2112930Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2113289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.2113384Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.2113389Z 2025-09-07T07:43:06.2113490Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.2113629Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2113916Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2113993Z return mod(**inputs) 2025-09-07T07:43:06.2114362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2114451Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2114821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2114909Z layer_outputs = layer_module( 2025-09-07T07:43:06.2115252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2115351Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2115799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2115922Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2116336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.2116463Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.2116827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.2116919Z return self.weight * hidden_states 2025-09-07T07:43:06.2116924Z 2025-09-07T07:43:06.2117059Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2117390Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2117469Z return mod(**inputs) 2025-09-07T07:43:06.2117839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2117930Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2118299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2118386Z layer_outputs = layer_module( 2025-09-07T07:43:06.2118729Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2118825Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2119186Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2119302Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2119663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.2119818Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.2120182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.2120283Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.2120292Z 2025-09-07T07:43:06.2120427Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2120712Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2120793Z return mod(**inputs) 2025-09-07T07:43:06.2121159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2121251Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2121619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2121704Z layer_outputs = layer_module( 2025-09-07T07:43:06.2122051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2122147Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2122515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2122628Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2122987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.2123141Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.2123504Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.2123610Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.2123615Z 2025-09-07T07:43:06.2123749Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2124124Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2124208Z return mod(**inputs) 2025-09-07T07:43:06.2124573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2124698Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2125067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2125158Z layer_outputs = layer_module( 2025-09-07T07:43:06.2125497Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2125593Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2125998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2126109Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2126480Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.2126632Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.2127002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.2127099Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.2127104Z 2025-09-07T07:43:06.2127238Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2127529Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2127605Z return mod(**inputs) 2025-09-07T07:43:06.2127979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2128069Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2128437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2128529Z layer_outputs = layer_module( 2025-09-07T07:43:06.2128871Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2128969Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2129333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2129444Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2129811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 343, in forward 2025-09-07T07:43:06.2129978Z hidden_states = hidden_states + self.dropout(forwarded_states) 2025-09-07T07:43:06.2129983Z 2025-09-07T07:43:06.2130083Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.2130219Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2130513Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2130590Z return mod(**inputs) 2025-09-07T07:43:06.2130958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2131049Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2131414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2131505Z layer_outputs = layer_module( 2025-09-07T07:43:06.2131844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2131942Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2132311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2132494Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2132863Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.2133035Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.2133399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.2133518Z return self.weight * hidden_states 2025-09-07T07:43:06.2133523Z 2025-09-07T07:43:06.2133660Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2133951Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2134070Z return mod(**inputs) 2025-09-07T07:43:06.2134439Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2134528Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2134900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2134988Z layer_outputs = layer_module( 2025-09-07T07:43:06.2135333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2135432Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2135796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2135896Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2136262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2136368Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2136736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.2136831Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.2136835Z 2025-09-07T07:43:06.2136973Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2137262Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2137337Z return mod(**inputs) 2025-09-07T07:43:06.2137708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2137797Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2138167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2138252Z layer_outputs = layer_module( 2025-09-07T07:43:06.2138595Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2138696Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2139062Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2139167Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2139532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2139636Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2139999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.2140089Z key_states = self.k(current_states) 2025-09-07T07:43:06.2140093Z 2025-09-07T07:43:06.2140232Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2140521Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2140601Z return mod(**inputs) 2025-09-07T07:43:06.2141060Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2141154Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2141521Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2141643Z layer_outputs = layer_module( 2025-09-07T07:43:06.2141984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2142082Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2142444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2142546Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2142951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2143059Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2143424Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2143593Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2143603Z 2025-09-07T07:43:06.2143736Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2144022Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2144101Z return mod(**inputs) 2025-09-07T07:43:06.2144469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2144560Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2144931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2145017Z layer_outputs = layer_module( 2025-09-07T07:43:06.2145365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2145462Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2145832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2145931Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2146295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2146402Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2146763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2146937Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2146942Z 2025-09-07T07:43:06.2147075Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2147366Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2147442Z return mod(**inputs) 2025-09-07T07:43:06.2147808Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2147902Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2148268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2148356Z layer_outputs = layer_module( 2025-09-07T07:43:06.2148697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2148792Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2149164Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2149265Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2149742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2149852Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2150252Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2150423Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2150428Z 2025-09-07T07:43:06.2150563Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2150857Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2150937Z return mod(**inputs) 2025-09-07T07:43:06.2151360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2151460Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2151830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2151921Z layer_outputs = layer_module( 2025-09-07T07:43:06.2152263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2152368Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2152730Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2152830Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2153197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2153305Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2153673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.2153887Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.2153893Z 2025-09-07T07:43:06.2154033Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2154321Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2154397Z return mod(**inputs) 2025-09-07T07:43:06.2154769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2154858Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2155230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2155316Z layer_outputs = layer_module( 2025-09-07T07:43:06.2155656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2155756Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2156123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2156226Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2156589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2156694Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2157055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.2157148Z value_states = self.v(current_states) 2025-09-07T07:43:06.2157153Z 2025-09-07T07:43:06.2157293Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2157580Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2157659Z return mod(**inputs) 2025-09-07T07:43:06.2158121Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2158219Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2158594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2158714Z layer_outputs = layer_module( 2025-09-07T07:43:06.2159059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2159157Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2159519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2159625Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2160011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2160119Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2160484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2160627Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2160634Z 2025-09-07T07:43:06.2160769Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2161056Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2161134Z return mod(**inputs) 2025-09-07T07:43:06.2161502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2161595Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2161962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2162049Z layer_outputs = layer_module( 2025-09-07T07:43:06.2162398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2162495Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2162862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2162965Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2163328Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2163434Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2163796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2163943Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2163947Z 2025-09-07T07:43:06.2164083Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2164373Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2164452Z return mod(**inputs) 2025-09-07T07:43:06.2164818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2164911Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2165361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2165455Z layer_outputs = layer_module( 2025-09-07T07:43:06.2165795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2165891Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2166259Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2166360Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2166880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2166991Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2167399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2167540Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2167545Z 2025-09-07T07:43:06.2167679Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2167968Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2168044Z return mod(**inputs) 2025-09-07T07:43:06.2168415Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2168580Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2168954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2169043Z layer_outputs = layer_module( 2025-09-07T07:43:06.2169383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2169487Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2169851Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2169951Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2170316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2170420Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2170787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.2170926Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.2170931Z 2025-09-07T07:43:06.2171073Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2171358Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2171438Z return mod(**inputs) 2025-09-07T07:43:06.2171807Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2171897Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2172269Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2172354Z layer_outputs = layer_module( 2025-09-07T07:43:06.2172694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2172797Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2173161Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2173269Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2173628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2173738Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2174100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.2174190Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.2174195Z 2025-09-07T07:43:06.2174296Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.2174431Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2174724Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2174801Z return mod(**inputs) 2025-09-07T07:43:06.2175263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2175361Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2175733Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2175858Z layer_outputs = layer_module( 2025-09-07T07:43:06.2176199Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2176296Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2176661Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2176764Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2177168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:43:06.2177305Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.2177677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.2177770Z return self.weight * hidden_states 2025-09-07T07:43:06.2177777Z 2025-09-07T07:43:06.2177912Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2178201Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2178277Z return mod(**inputs) 2025-09-07T07:43:06.2178646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2178736Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2179101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2179192Z layer_outputs = layer_module( 2025-09-07T07:43:06.2179536Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2179638Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2179999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2180104Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2180468Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2180575Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2180943Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.2181039Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.2181043Z 2025-09-07T07:43:06.2181180Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2181466Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2181545Z return mod(**inputs) 2025-09-07T07:43:06.2181916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2182008Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2182375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2182461Z layer_outputs = layer_module( 2025-09-07T07:43:06.2182802Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2182903Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2183264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2183374Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2183831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2183947Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2184309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.2184431Z key_states = self.k(current_states) 2025-09-07T07:43:06.2184436Z 2025-09-07T07:43:06.2184573Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2184859Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2184939Z return mod(**inputs) 2025-09-07T07:43:06.2185306Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2185426Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2185794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2185882Z layer_outputs = layer_module( 2025-09-07T07:43:06.2186223Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2186321Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2186685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2186790Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2187151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2187262Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2187626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2187799Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2187804Z 2025-09-07T07:43:06.2187943Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2188228Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2188309Z return mod(**inputs) 2025-09-07T07:43:06.2188677Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2188768Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2189136Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2189222Z layer_outputs = layer_module( 2025-09-07T07:43:06.2189564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2189663Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2190028Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2190133Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2190498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2190609Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2190971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2191145Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2191150Z 2025-09-07T07:43:06.2191286Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2191576Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2191655Z return mod(**inputs) 2025-09-07T07:43:06.2192022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2192211Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2192583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2192710Z layer_outputs = layer_module( 2025-09-07T07:43:06.2193057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2193153Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2193522Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2193625Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2193995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2194150Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2194521Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2194692Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2194699Z 2025-09-07T07:43:06.2194834Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2195124Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2195202Z return mod(**inputs) 2025-09-07T07:43:06.2195572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2195662Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2196029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2196121Z layer_outputs = layer_module( 2025-09-07T07:43:06.2196462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2196565Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2196929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2197039Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2197401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2197509Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2197874Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.2198084Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.2198091Z 2025-09-07T07:43:06.2198233Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2198518Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2198597Z return mod(**inputs) 2025-09-07T07:43:06.2198969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2199061Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2199432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2199519Z layer_outputs = layer_module( 2025-09-07T07:43:06.2199861Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2199959Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2200324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2200435Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2200924Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2201041Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2201408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.2201540Z value_states = self.v(current_states) 2025-09-07T07:43:06.2201545Z 2025-09-07T07:43:06.2201687Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2201975Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2202054Z return mod(**inputs) 2025-09-07T07:43:06.2202422Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2202542Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2202911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2202999Z layer_outputs = layer_module( 2025-09-07T07:43:06.2203344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2203445Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2203811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2203916Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2204279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2204392Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2204757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2204900Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2204905Z 2025-09-07T07:43:06.2205044Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2205331Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2205413Z return mod(**inputs) 2025-09-07T07:43:06.2205781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2205873Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2206240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2206325Z layer_outputs = layer_module( 2025-09-07T07:43:06.2206669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2206767Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2207131Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2207237Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2207603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2207712Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2208073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2208214Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2208220Z 2025-09-07T07:43:06.2208355Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2208642Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2208722Z return mod(**inputs) 2025-09-07T07:43:06.2209089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2209287Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2209660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2209771Z layer_outputs = layer_module( 2025-09-07T07:43:06.2210113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2210214Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2210577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2210679Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2211046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2211193Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2211565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2211704Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2211709Z 2025-09-07T07:43:06.2211847Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2212139Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2212216Z return mod(**inputs) 2025-09-07T07:43:06.2212587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2212676Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2213042Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2213133Z layer_outputs = layer_module( 2025-09-07T07:43:06.2213473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2213578Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2213940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2214050Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2214411Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2214516Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2214884Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.2215023Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.2215031Z 2025-09-07T07:43:06.2215170Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2215457Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2215537Z return mod(**inputs) 2025-09-07T07:43:06.2215913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2216006Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2216376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2216462Z layer_outputs = layer_module( 2025-09-07T07:43:06.2216805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2216903Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2217266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2217374Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2217735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2217927Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2218298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.2218420Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.2218425Z 2025-09-07T07:43:06.2218526Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.2218662Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2218950Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2219027Z return mod(**inputs) 2025-09-07T07:43:06.2219393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2219523Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2219891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2219986Z layer_outputs = layer_module( 2025-09-07T07:43:06.2220327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2220428Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2220791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2220903Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2221268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.2221393Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.2221762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.2221855Z return self.weight * hidden_states 2025-09-07T07:43:06.2221860Z 2025-09-07T07:43:06.2221996Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2222287Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2222365Z return mod(**inputs) 2025-09-07T07:43:06.2222736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2222825Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2223193Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2223284Z layer_outputs = layer_module( 2025-09-07T07:43:06.2223624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2223727Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2224089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2224208Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2224572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.2224727Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.2225094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.2225192Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.2225196Z 2025-09-07T07:43:06.2225337Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2225626Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2225705Z return mod(**inputs) 2025-09-07T07:43:06.2226076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2226236Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2226613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2226732Z layer_outputs = layer_module( 2025-09-07T07:43:06.2227078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2227174Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2227537Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2227655Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2228018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.2228207Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.2228572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.2228670Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.2228677Z 2025-09-07T07:43:06.2228814Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2229097Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2229176Z return mod(**inputs) 2025-09-07T07:43:06.2229542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2229630Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2229998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2230086Z layer_outputs = layer_module( 2025-09-07T07:43:06.2230427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2230528Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2230894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2231009Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2231372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.2231529Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.2231893Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.2231995Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.2232002Z 2025-09-07T07:43:06.2232098Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.2232233Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2232524Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2232610Z return mod(**inputs) 2025-09-07T07:43:06.2232983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2233075Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2233444Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2233528Z layer_outputs = layer_module( 2025-09-07T07:43:06.2233868Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2233969Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2234334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2234439Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2234904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.2235046Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.2235446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.2235540Z return self.weight * hidden_states 2025-09-07T07:43:06.2235545Z 2025-09-07T07:43:06.2235685Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2235970Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2236047Z return mod(**inputs) 2025-09-07T07:43:06.2236416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2236545Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2236917Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2237006Z layer_outputs = layer_module( 2025-09-07T07:43:06.2237353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2237452Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2237815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2237919Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2238281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2238389Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2238752Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.2238844Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.2238849Z 2025-09-07T07:43:06.2238989Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2239275Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2239357Z return mod(**inputs) 2025-09-07T07:43:06.2239723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2239814Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2240180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2240265Z layer_outputs = layer_module( 2025-09-07T07:43:06.2240611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2240709Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2241078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2241177Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2241537Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2241645Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2242006Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.2242100Z key_states = self.k(current_states) 2025-09-07T07:43:06.2242105Z 2025-09-07T07:43:06.2242238Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2242522Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2242606Z return mod(**inputs) 2025-09-07T07:43:06.2242971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2243147Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2243520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2243663Z layer_outputs = layer_module( 2025-09-07T07:43:06.2244006Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2244103Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2244471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2244572Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2244938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2245077Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2245445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2245617Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2245624Z 2025-09-07T07:43:06.2245758Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2246048Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2246123Z return mod(**inputs) 2025-09-07T07:43:06.2246486Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2246579Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2246944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2247034Z layer_outputs = layer_module( 2025-09-07T07:43:06.2247378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2247480Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2247842Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2247944Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2248305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2248408Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2248772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2248941Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2248948Z 2025-09-07T07:43:06.2249081Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2249371Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2249451Z return mod(**inputs) 2025-09-07T07:43:06.2249819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2249910Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2250282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2250368Z layer_outputs = layer_module( 2025-09-07T07:43:06.2250711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2250813Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2251178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2251286Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2251743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2251852Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2252222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2252428Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2252432Z 2025-09-07T07:43:06.2252571Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2252858Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2252939Z return mod(**inputs) 2025-09-07T07:43:06.2253305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2253428Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2253799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2253888Z layer_outputs = layer_module( 2025-09-07T07:43:06.2254232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2254332Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2254696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2254801Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2255164Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2255272Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2255635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.2255848Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.2255857Z 2025-09-07T07:43:06.2255993Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2256278Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2256359Z return mod(**inputs) 2025-09-07T07:43:06.2256727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2256818Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2257184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2257270Z layer_outputs = layer_module( 2025-09-07T07:43:06.2257614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2257714Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2258082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2258186Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2258548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2258657Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2259020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.2259117Z value_states = self.v(current_states) 2025-09-07T07:43:06.2259121Z 2025-09-07T07:43:06.2259257Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2259549Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2259627Z return mod(**inputs) 2025-09-07T07:43:06.2259994Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2260162Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2260539Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2260658Z layer_outputs = layer_module( 2025-09-07T07:43:06.2260999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2261096Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2261463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2261569Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2261937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2262087Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2262454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2262598Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2262603Z 2025-09-07T07:43:06.2262741Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2263029Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2263107Z return mod(**inputs) 2025-09-07T07:43:06.2263478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2263568Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2263931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2264022Z layer_outputs = layer_module( 2025-09-07T07:43:06.2264362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2264464Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2264827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2264930Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2265377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2265481Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2265849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2265989Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2265996Z 2025-09-07T07:43:06.2266135Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2266419Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2266496Z return mod(**inputs) 2025-09-07T07:43:06.2266869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2266961Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2267331Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2267417Z layer_outputs = layer_module( 2025-09-07T07:43:06.2267760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2267861Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2268225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2268331Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2268696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2268964Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2269338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2269532Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2269537Z 2025-09-07T07:43:06.2269673Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2269959Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2270039Z return mod(**inputs) 2025-09-07T07:43:06.2270403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2270547Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2270915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2271003Z layer_outputs = layer_module( 2025-09-07T07:43:06.2271349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2271450Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2271813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2271916Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2272277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2272383Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2272744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.2272884Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.2272893Z 2025-09-07T07:43:06.2273027Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2273314Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2273396Z return mod(**inputs) 2025-09-07T07:43:06.2273766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2273858Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2274225Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2274311Z layer_outputs = layer_module( 2025-09-07T07:43:06.2274657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2274753Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2275119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2275223Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2275586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2275695Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2276058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.2276152Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.2276156Z 2025-09-07T07:43:06.2276291Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2276578Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2276654Z return mod(**inputs) 2025-09-07T07:43:06.2277023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2277116Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2277593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2277688Z layer_outputs = layer_module( 2025-09-07T07:43:06.2278065Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2278161Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2278526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2278626Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2278992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T07:43:06.2279204Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T07:43:06.2279208Z 2025-09-07T07:43:06.2279312Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.2279448Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2279735Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2279819Z return mod(**inputs) 2025-09-07T07:43:06.2280188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2280281Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2280647Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2280732Z layer_outputs = layer_module( 2025-09-07T07:43:06.2281081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2281180Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2281544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2281649Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2282012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:43:06.2282152Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.2282513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.2282608Z return self.weight * hidden_states 2025-09-07T07:43:06.2282613Z 2025-09-07T07:43:06.2282747Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2283037Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2283114Z return mod(**inputs) 2025-09-07T07:43:06.2283482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2283574Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2283942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2284031Z layer_outputs = layer_module( 2025-09-07T07:43:06.2284370Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2284468Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2284835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2284938Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2285303Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2285416Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2285776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.2285960Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.2285966Z 2025-09-07T07:43:06.2286107Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2286429Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2286505Z return mod(**inputs) 2025-09-07T07:43:06.2286876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2286965Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2287333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2287459Z layer_outputs = layer_module( 2025-09-07T07:43:06.2287800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2287899Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2288265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2288371Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2288734Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2288842Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2289207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.2289298Z key_states = self.k(current_states) 2025-09-07T07:43:06.2289303Z 2025-09-07T07:43:06.2289440Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2289727Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2289805Z return mod(**inputs) 2025-09-07T07:43:06.2290182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2290271Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2290643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2290728Z layer_outputs = layer_module( 2025-09-07T07:43:06.2291066Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2291170Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2291536Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2291644Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2292007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2292116Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2292487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2292657Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2292661Z 2025-09-07T07:43:06.2292797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2293086Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2293165Z return mod(**inputs) 2025-09-07T07:43:06.2293533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2293622Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2293993Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2294079Z layer_outputs = layer_module( 2025-09-07T07:43:06.2294512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2294613Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2295012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2295121Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2295481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2295592Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2295952Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2296160Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2296169Z 2025-09-07T07:43:06.2296306Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2296598Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2296684Z return mod(**inputs) 2025-09-07T07:43:06.2297052Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2297148Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2297514Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2297598Z layer_outputs = layer_module( 2025-09-07T07:43:06.2297942Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2298043Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2298410Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2298514Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2298879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2298993Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2299354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2299525Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2299530Z 2025-09-07T07:43:06.2299665Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2299954Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2300031Z return mod(**inputs) 2025-09-07T07:43:06.2300403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2300494Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2300865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2300955Z layer_outputs = layer_module( 2025-09-07T07:43:06.2301299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2301396Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2301766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2301870Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2302236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2302347Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2302710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.2303016Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.2303022Z 2025-09-07T07:43:06.2303163Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2303489Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2303565Z return mod(**inputs) 2025-09-07T07:43:06.2303936Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2304027Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2304391Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2304505Z layer_outputs = layer_module( 2025-09-07T07:43:06.2304845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2304945Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2305310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2305416Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2305784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2305893Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2306261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.2306355Z value_states = self.v(current_states) 2025-09-07T07:43:06.2306360Z 2025-09-07T07:43:06.2306499Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2306785Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2306861Z return mod(**inputs) 2025-09-07T07:43:06.2307235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2307325Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2307695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2307780Z layer_outputs = layer_module( 2025-09-07T07:43:06.2308119Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2308219Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2308584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2308693Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2309056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2309161Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2309527Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2309667Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2309672Z 2025-09-07T07:43:06.2309810Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2310096Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2310175Z return mod(**inputs) 2025-09-07T07:43:06.2310543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2310633Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2311007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2311093Z layer_outputs = layer_module( 2025-09-07T07:43:06.2311553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2311654Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2312053Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2312161Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2312524Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2312635Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2312999Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2313174Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2313179Z 2025-09-07T07:43:06.2313313Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2313601Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2313682Z return mod(**inputs) 2025-09-07T07:43:06.2314047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2318180Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2318634Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2318725Z layer_outputs = layer_module( 2025-09-07T07:43:06.2319075Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2319173Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2319549Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2319651Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2320015Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2320127Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2320491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2320632Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2320638Z 2025-09-07T07:43:06.2320773Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2321062Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2321138Z return mod(**inputs) 2025-09-07T07:43:06.2321512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2321602Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2321973Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2322061Z layer_outputs = layer_module( 2025-09-07T07:43:06.2322401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2322505Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2322865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2322971Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2323331Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2323440Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2323804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.2324093Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.2324099Z 2025-09-07T07:43:06.2324250Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2324535Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2324655Z return mod(**inputs) 2025-09-07T07:43:06.2325022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2325110Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2325483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2325568Z layer_outputs = layer_module( 2025-09-07T07:43:06.2325948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2326045Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2326408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2326515Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2326878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2326986Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2327347Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.2327437Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.2327445Z 2025-09-07T07:43:06.2327541Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.2327678Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2327969Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2328043Z return mod(**inputs) 2025-09-07T07:43:06.2328418Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2328507Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2328874Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2328961Z layer_outputs = layer_module( 2025-09-07T07:43:06.2329303Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2329402Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2329767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2329880Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2330246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.2330376Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.2330741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.2330839Z return self.weight * hidden_states 2025-09-07T07:43:06.2330843Z 2025-09-07T07:43:06.2330975Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2331266Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2331343Z return mod(**inputs) 2025-09-07T07:43:06.2331711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2331801Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2332170Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2332256Z layer_outputs = layer_module( 2025-09-07T07:43:06.2332692Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2332800Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2333201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2333317Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2333679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.2333831Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.2334195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.2334360Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.2334366Z 2025-09-07T07:43:06.2334504Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2334794Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2334874Z return mod(**inputs) 2025-09-07T07:43:06.2335239Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2335332Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2335707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2335792Z layer_outputs = layer_module( 2025-09-07T07:43:06.2336139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2336237Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2336600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2336715Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2337081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.2337239Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.2337601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.2337700Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.2337708Z 2025-09-07T07:43:06.2337846Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2338130Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2338209Z return mod(**inputs) 2025-09-07T07:43:06.2338578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2338671Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2339040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2339124Z layer_outputs = layer_module( 2025-09-07T07:43:06.2339472Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2339571Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2339939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2340051Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2340414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.2340570Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.2340931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.2341118Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.2341124Z 2025-09-07T07:43:06.2341228Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.2341400Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2341686Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2341762Z return mod(**inputs) 2025-09-07T07:43:06.2342132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2342219Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2342587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2342726Z layer_outputs = layer_module( 2025-09-07T07:43:06.2343071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2343175Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2343538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2343645Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2344007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T07:43:06.2344141Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.2344506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.2344599Z return self.weight * hidden_states 2025-09-07T07:43:06.2344606Z 2025-09-07T07:43:06.2344742Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2345028Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2345107Z return mod(**inputs) 2025-09-07T07:43:06.2345477Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2345564Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2345938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2346023Z layer_outputs = layer_module( 2025-09-07T07:43:06.2346369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2346465Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2346830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2346936Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2347298Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2347407Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2347767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.2347863Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.2347871Z 2025-09-07T07:43:06.2348005Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2348292Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2348371Z return mod(**inputs) 2025-09-07T07:43:06.2348735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2348829Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2349193Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2349278Z layer_outputs = layer_module( 2025-09-07T07:43:06.2349704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2349805Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2350207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2350307Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2350667Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2350774Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2351135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.2351259Z key_states = self.k(current_states) 2025-09-07T07:43:06.2351265Z 2025-09-07T07:43:06.2351400Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2351693Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2351771Z return mod(**inputs) 2025-09-07T07:43:06.2352136Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2352230Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2352593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2352683Z layer_outputs = layer_module( 2025-09-07T07:43:06.2353023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2353117Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2353483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2353581Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2353950Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2354052Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2354416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2354590Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2354594Z 2025-09-07T07:43:06.2354727Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2355018Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2355096Z return mod(**inputs) 2025-09-07T07:43:06.2355469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2355559Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2355929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2356018Z layer_outputs = layer_module( 2025-09-07T07:43:06.2356361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2356463Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2356826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2356926Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2357291Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2357398Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2357763Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2358027Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2358033Z 2025-09-07T07:43:06.2358176Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2358495Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2358571Z return mod(**inputs) 2025-09-07T07:43:06.2358941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2359031Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2359400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2359485Z layer_outputs = layer_module( 2025-09-07T07:43:06.2359868Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2359969Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2360336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2360440Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2360803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2360905Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2361267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2361436Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2361441Z 2025-09-07T07:43:06.2361580Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2361871Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2361947Z return mod(**inputs) 2025-09-07T07:43:06.2362319Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2362406Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2362781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2362867Z layer_outputs = layer_module( 2025-09-07T07:43:06.2363208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2363303Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2363663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2363769Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2364129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2364237Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2364602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.2364818Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.2364823Z 2025-09-07T07:43:06.2364957Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2365242Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2365394Z return mod(**inputs) 2025-09-07T07:43:06.2365762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2365851Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2366219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2366304Z layer_outputs = layer_module( 2025-09-07T07:43:06.2366822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2366926Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2367344Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2367444Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2367803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2367909Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2368271Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.2368424Z value_states = self.v(current_states) 2025-09-07T07:43:06.2368429Z 2025-09-07T07:43:06.2368563Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2368854Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2368929Z return mod(**inputs) 2025-09-07T07:43:06.2369296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2369390Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2369754Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2369842Z layer_outputs = layer_module( 2025-09-07T07:43:06.2370184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2370279Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2370649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2370746Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2371114Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2371217Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2371585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2371723Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2371729Z 2025-09-07T07:43:06.2371863Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2372153Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2372229Z return mod(**inputs) 2025-09-07T07:43:06.2372601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2372689Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2373055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2373143Z layer_outputs = layer_module( 2025-09-07T07:43:06.2373484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2373587Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2373949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2374049Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2374414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2374518Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2374889Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2375027Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2375111Z 2025-09-07T07:43:06.2375257Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2375545Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2375668Z return mod(**inputs) 2025-09-07T07:43:06.2376039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2376127Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2376499Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2376589Z layer_outputs = layer_module( 2025-09-07T07:43:06.2376968Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2377068Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2377434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2377538Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2377900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2378004Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2378364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2378499Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2378504Z 2025-09-07T07:43:06.2378641Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2378929Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2379010Z return mod(**inputs) 2025-09-07T07:43:06.2379378Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2379470Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2379834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2379922Z layer_outputs = layer_module( 2025-09-07T07:43:06.2380262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2380357Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2380718Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2380821Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2381187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2381295Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2381658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.2381800Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.2381807Z 2025-09-07T07:43:06.2381942Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2382227Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2382309Z return mod(**inputs) 2025-09-07T07:43:06.2382676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2382767Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2383135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2383224Z layer_outputs = layer_module( 2025-09-07T07:43:06.2383659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2383759Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2384125Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T07:43:06.2384261Z self_attention_outputs = self.layer[0]( 2025-09-07T07:43:06.2384626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T07:43:06.2384728Z attention_output = self.SelfAttention( 2025-09-07T07:43:06.2385091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.2385217Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.2385222Z 2025-09-07T07:43:06.2385317Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.2385456Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2385745Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2385822Z return mod(**inputs) 2025-09-07T07:43:06.2386192Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2386285Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2386656Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2386742Z layer_outputs = layer_module( 2025-09-07T07:43:06.2387084Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2387183Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2387548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2387654Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2388019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T07:43:06.2388160Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.2388521Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.2388612Z return self.weight * hidden_states 2025-09-07T07:43:06.2388617Z 2025-09-07T07:43:06.2388753Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2389039Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2389118Z return mod(**inputs) 2025-09-07T07:43:06.2389485Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2389573Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2389944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2390032Z layer_outputs = layer_module( 2025-09-07T07:43:06.2390374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2390473Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2390836Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2390941Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2391305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2391419Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2391781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T07:43:06.2391879Z query_states = self.q(hidden_states) 2025-09-07T07:43:06.2391968Z 2025-09-07T07:43:06.2392113Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2392399Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2392511Z return mod(**inputs) 2025-09-07T07:43:06.2392879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2392970Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2393336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2393419Z layer_outputs = layer_module( 2025-09-07T07:43:06.2393785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2393882Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2394253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2394354Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2394717Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2394829Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2395189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T07:43:06.2395285Z key_states = self.k(current_states) 2025-09-07T07:43:06.2395290Z 2025-09-07T07:43:06.2395424Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2395715Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2395792Z return mod(**inputs) 2025-09-07T07:43:06.2396160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2396258Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2396625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2396715Z layer_outputs = layer_module( 2025-09-07T07:43:06.2397056Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2397151Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2397519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2397622Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2397988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2398096Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2398464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2398632Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2398639Z 2025-09-07T07:43:06.2398775Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2399063Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2399151Z return mod(**inputs) 2025-09-07T07:43:06.2399517Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2399611Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2399977Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2400065Z layer_outputs = layer_module( 2025-09-07T07:43:06.2400410Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2400572Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2400949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2401077Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2401443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2401550Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2401913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2402085Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2402112Z 2025-09-07T07:43:06.2402250Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2402540Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2402619Z return mod(**inputs) 2025-09-07T07:43:06.2402986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2403082Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2403449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2403539Z layer_outputs = layer_module( 2025-09-07T07:43:06.2403880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2403979Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2404343Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2404452Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2404821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2404928Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2405293Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T07:43:06.2405464Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T07:43:06.2405469Z 2025-09-07T07:43:06.2405603Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2405892Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2405969Z return mod(**inputs) 2025-09-07T07:43:06.2406338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2406431Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2406799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2406889Z layer_outputs = layer_module( 2025-09-07T07:43:06.2407231Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2407335Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2407696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2407800Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2408160Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2408267Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2408643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T07:43:06.2408854Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T07:43:06.2408859Z 2025-09-07T07:43:06.2409106Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2409401Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2409511Z return mod(**inputs) 2025-09-07T07:43:06.2409882Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2409971Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2410342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2410428Z layer_outputs = layer_module( 2025-09-07T07:43:06.2410802Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2410899Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2411261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2411365Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2411726Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2411836Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2412201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T07:43:06.2412293Z value_states = self.v(current_states) 2025-09-07T07:43:06.2412297Z 2025-09-07T07:43:06.2412434Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2412718Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2412800Z return mod(**inputs) 2025-09-07T07:43:06.2413168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2413261Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2413632Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2413719Z layer_outputs = layer_module( 2025-09-07T07:43:06.2414063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2414159Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2414523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2414625Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2414988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2415099Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2415463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2415605Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2415612Z 2025-09-07T07:43:06.2415745Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2416030Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2416110Z return mod(**inputs) 2025-09-07T07:43:06.2416478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2416570Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2416937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2417029Z layer_outputs = layer_module( 2025-09-07T07:43:06.2417371Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2417564Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2417941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2418077Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2418445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2418553Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2418915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2419058Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2419104Z 2025-09-07T07:43:06.2419240Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2419528Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2419608Z return mod(**inputs) 2025-09-07T07:43:06.2419975Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2420070Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2420436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2420524Z layer_outputs = layer_module( 2025-09-07T07:43:06.2420865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2420964Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2421328Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2421433Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2421802Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2421909Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2422273Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T07:43:06.2422416Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T07:43:06.2422420Z 2025-09-07T07:43:06.2422554Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2422841Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2422915Z return mod(**inputs) 2025-09-07T07:43:06.2423282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2423374Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2423740Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2423828Z layer_outputs = layer_module( 2025-09-07T07:43:06.2424166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2424271Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2424633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2424738Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2425102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2425211Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2425575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T07:43:06.2425716Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T07:43:06.2425721Z 2025-09-07T07:43:06.2425945Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2426239Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2426348Z return mod(**inputs) 2025-09-07T07:43:06.2426719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2426808Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2427176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2427261Z layer_outputs = layer_module( 2025-09-07T07:43:06.2427607Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2427738Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2428102Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2428210Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2428572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T07:43:06.2428683Z attention_output = self.EncDecAttention( 2025-09-07T07:43:06.2429043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T07:43:06.2429132Z attn_output = self.o(attn_output) 2025-09-07T07:43:06.2429136Z 2025-09-07T07:43:06.2429273Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2429557Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2429637Z return mod(**inputs) 2025-09-07T07:43:06.2430006Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2430101Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2430470Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2430558Z layer_outputs = layer_module( 2025-09-07T07:43:06.2430902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2430998Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2431361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T07:43:06.2431462Z cross_attention_outputs = self.layer[1]( 2025-09-07T07:43:06.2431823Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 647, in forward 2025-09-07T07:43:06.2431998Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-09-07T07:43:06.2432003Z 2025-09-07T07:43:06.2432101Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.2432242Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2432527Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2432607Z return mod(**inputs) 2025-09-07T07:43:06.2432979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2433069Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2433437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2433523Z layer_outputs = layer_module( 2025-09-07T07:43:06.2433865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2433966Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2434413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2434533Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2434895Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T07:43:06.2435058Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T07:43:06.2435421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T07:43:06.2435512Z return self.weight * hidden_states 2025-09-07T07:43:06.2435517Z 2025-09-07T07:43:06.2435654Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2435939Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2436054Z return mod(**inputs) 2025-09-07T07:43:06.2436418Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2436510Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2436879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2436967Z layer_outputs = layer_module( 2025-09-07T07:43:06.2437309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2437407Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2437773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2437885Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2438248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.2438402Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.2438766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T07:43:06.2438866Z hidden_states = self.wi(hidden_states) 2025-09-07T07:43:06.2438874Z 2025-09-07T07:43:06.2439007Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2439290Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2439369Z return mod(**inputs) 2025-09-07T07:43:06.2439739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2439833Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2440197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2440286Z layer_outputs = layer_module( 2025-09-07T07:43:06.2440627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2440728Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2441096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2441209Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2441573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.2441726Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.2442088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T07:43:06.2442189Z hidden_states = self.act(hidden_states) 2025-09-07T07:43:06.2442195Z 2025-09-07T07:43:06.2442328Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2442616Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2442806Z return mod(**inputs) 2025-09-07T07:43:06.2443182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T07:43:06.2443309Z decoder_outputs = self.decoder( 2025-09-07T07:43:06.2443676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T07:43:06.2443763Z layer_outputs = layer_module( 2025-09-07T07:43:06.2444101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T07:43:06.2444200Z return super().__call__(*args, **kwargs) 2025-09-07T07:43:06.2444561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T07:43:06.2444706Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T07:43:06.2445073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T07:43:06.2445227Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T07:43:06.2445594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T07:43:06.2445694Z hidden_states = self.wo(hidden_states) 2025-09-07T07:43:06.2445699Z 2025-09-07T07:43:06.2445794Z cudagraph partition due to non gpu ops 2025-09-07T07:43:06.2445931Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2446216Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2446295Z return mod(**inputs) 2025-09-07T07:43:06.2446665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1789, in forward 2025-09-07T07:43:06.2446825Z sequence_output = sequence_output * (self.model_dim**-0.5) 2025-09-07T07:43:06.2446830Z 2025-09-07T07:43:06.2446967Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T07:43:06.2447252Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T07:43:06.2447334Z return mod(**inputs) 2025-09-07T07:43:06.2447700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/t5/modeling_t5.py", line 1791, in forward 2025-09-07T07:43:06.2447806Z lm_logits = self.lm_head(sequence_output) 2025-09-07T07:43:06.2447810Z 2025-09-07T07:43:29.1660694Z pass 2025-09-07T07:43:29.1661165Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:43:34.3214656Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:43:34.3215870Z import pynvml # type: ignore[import] 2025-09-07T07:43:36.9135577Z 2025-09-07T07:43:51.4154595Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:43:51.4154892Z loading model: 0it [00:14, ?it/s] 2025-09-07T07:43:51.4812634Z cpu eval hf_T5_large 2025-09-07T07:43:51.5103907Z pass_due_to_skip 2025-09-07T07:43:51.5106553Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:43:53.0775373Z accuracy pass_rate=80.00% 2025-09-07T07:43:53.0780806Z calls_captured gmean=0.00x mean=637.600x 2025-09-07T07:43:53.0784658Z unique_graphs gmean=0.00x mean=2.000x 2025-09-07T07:43:53.0789103Z graph_breaks gmean=0.00x mean=1.600x 2025-09-07T07:43:53.0793002Z unique_graph_breaks gmean=0.00x mean=0.800x 2025-09-07T07:43:53.0796785Z autograd_captures gmean=0.00x mean=0.000x 2025-09-07T07:43:53.0800480Z autograd_compiles gmean=0.00x mean=0.000x 2025-09-07T07:43:53.0804174Z cudagraph_skips gmean=0.00x mean=2.000x 2025-09-07T07:43:53.0805631Z compilation_latency mean=20.980 seconds 2025-09-07T07:43:53.5421654Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *cppwrapper-true* ]] 2025-09-07T07:43:53.5422461Z + TORCHINDUCTOR_CPP_WRAPPER=1 2025-09-07T07:43:53.5423868Z + taskset -c 0-62 python benchmarks/dynamo/torchbench.py --accuracy --no-translation-validation --inference --bfloat16 --backend inductor --disable-cudagraphs --device cpu --total-partitions 15 --partition-id 7 --output /var/lib/jenkins/workspace/test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_aarch64_accuracy.csv 2025-09-07T07:43:53.8755475Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:43:53.8756965Z import pynvml # type: ignore[import] 2025-09-07T07:43:56.7825321Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:43:56.7826485Z import pynvml # type: ignore[import] 2025-09-07T07:43:59.3441629Z 2025-09-07T07:44:01.1541041Z loading model: 0it [00:00, ?it/s]Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T07:44:01.1542905Z WARNING:transformers.models.reformer.modeling_reformer:Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T07:44:01.3334092Z 2025-09-07T07:44:01.3334205Z loading model: 0it [00:01, ?it/s] 2025-09-07T07:44:01.3372007Z cpu eval hf_Reformer 2025-09-07T07:44:03.2728512Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:44:03.5952729Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:44:03.9146532Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:44:10.8137859Z W0907 07:44:10.813000 13119 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] Graph break from `Tensor.item()`, consider setting: 2025-09-07T07:44:10.8138843Z W0907 07:44:10.813000 13119 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] torch._dynamo.config.capture_scalar_outputs = True 2025-09-07T07:44:10.8139576Z W0907 07:44:10.813000 13119 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] or: 2025-09-07T07:44:10.8140308Z W0907 07:44:10.813000 13119 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1 2025-09-07T07:44:10.8141177Z W0907 07:44:10.813000 13119 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] to include these operations in the captured graph. 2025-09-07T07:44:10.8141889Z W0907 07:44:10.813000 13119 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T07:44:10.8142544Z W0907 07:44:10.813000 13119 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] Graph break: from user code at: 2025-09-07T07:44:10.8143817Z W0907 07:44:10.813000 13119 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 2132, in torch_dynamo_resume_in_forward_at_2104 2025-09-07T07:44:10.8145111Z W0907 07:44:10.813000 13119 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] embedding_output = self.embeddings( 2025-09-07T07:44:10.8146577Z W0907 07:44:10.813000 13119 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 369, in forward 2025-09-07T07:44:10.8147951Z W0907 07:44:10.813000 13119 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] position_embeddings = self.position_embeddings(position_ids) 2025-09-07T07:44:10.8149181Z W0907 07:44:10.813000 13119 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 293, in forward 2025-09-07T07:44:10.8150346Z W0907 07:44:10.813000 13119 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] max_position_id = position_ids.max().item() 2025-09-07T07:44:10.8151139Z W0907 07:44:10.813000 13119 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T07:44:10.8151686Z W0907 07:44:10.813000 13119 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T07:44:32.0663637Z pass 2025-09-07T07:44:32.0664135Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:44:33.7881521Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:44:33.7882794Z import pynvml # type: ignore[import] 2025-09-07T07:44:36.3573689Z 2025-09-07T07:44:46.3752924Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:44:46.3753238Z loading model: 0it [00:10, ?it/s] 2025-09-07T07:44:46.3975544Z cpu eval hf_Roberta_base 2025-09-07T07:44:52.2153276Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:44:53.1796378Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:44:54.1300882Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:45:14.6976596Z pass 2025-09-07T07:45:14.6977052Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:45:17.4040535Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:45:17.4041717Z import pynvml # type: ignore[import] 2025-09-07T07:45:20.0054567Z 2025-09-07T07:45:23.3948407Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:45:23.3948693Z loading model: 0it [00:03, ?it/s] 2025-09-07T07:45:23.4045635Z cpu eval hf_T5 2025-09-07T07:45:25.1059780Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:45:25.5320044Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:45:25.9548838Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:45:56.0813177Z pass 2025-09-07T07:45:56.0813656Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:45:58.0627896Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:45:58.0629105Z import pynvml # type: ignore[import] 2025-09-07T07:46:00.6633151Z 2025-09-07T07:46:12.4778201Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:46:12.4779353Z loading model: 0it [00:11, ?it/s] 2025-09-07T07:46:12.5022929Z cpu eval hf_T5_base 2025-09-07T07:46:47.2210840Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:46:56.0029173Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:47:04.7756635Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:48:14.8958705Z pass 2025-09-07T07:48:14.8959178Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:48:20.1364520Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:48:20.1366139Z import pynvml # type: ignore[import] 2025-09-07T07:48:22.7091016Z 2025-09-07T07:48:37.2646810Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:48:37.2647112Z loading model: 0it [00:14, ?it/s] 2025-09-07T07:48:37.3295256Z cpu eval hf_T5_large 2025-09-07T07:48:37.3587901Z pass_due_to_skip 2025-09-07T07:48:37.3590685Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:48:38.9449709Z accuracy pass_rate=80.00% 2025-09-07T07:48:38.9454890Z calls_captured gmean=0.00x mean=637.000x 2025-09-07T07:48:38.9458837Z unique_graphs gmean=0.00x mean=2.000x 2025-09-07T07:48:38.9463184Z graph_breaks gmean=0.00x mean=1.600x 2025-09-07T07:48:38.9467124Z unique_graph_breaks gmean=0.00x mean=0.800x 2025-09-07T07:48:38.9470796Z autograd_captures gmean=0.00x mean=0.000x 2025-09-07T07:48:38.9474427Z autograd_compiles gmean=0.00x mean=0.000x 2025-09-07T07:48:38.9478050Z cudagraph_skips gmean=0.00x mean=0.000x 2025-09-07T07:48:38.9479162Z compilation_latency mean=26.742 seconds 2025-09-07T07:48:39.3954665Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *freezing_cudagraphs-true* ]] 2025-09-07T07:48:39.3955847Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *freeze_autotune_cudagraphs-true* ]] 2025-09-07T07:48:39.3956918Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *aotinductor-true* ]] 2025-09-07T07:48:39.3957559Z + [[ inference == \i\n\f\e\r\e\n\c\e ]] 2025-09-07T07:48:39.3957822Z + [[ accuracy == \a\c\c\u\r\a\c\y ]] 2025-09-07T07:48:39.3959149Z + taskset -c 0-62 python benchmarks/dynamo/torchbench.py --accuracy --no-translation-validation --inference --bfloat16 --export --disable-cudagraphs --device cpu --total-partitions 15 --partition-id 7 --output /var/lib/jenkins/workspace/test/test-reports/inductor_export_torchbench_bfloat16_inference_cpu_aarch64_accuracy.csv 2025-09-07T07:48:39.7295777Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:48:39.7296942Z import pynvml # type: ignore[import] 2025-09-07T07:48:42.6421104Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:48:42.6422259Z import pynvml # type: ignore[import] 2025-09-07T07:48:45.2178448Z 2025-09-07T07:48:47.0307829Z loading model: 0it [00:00, ?it/s]Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T07:48:47.0309764Z WARNING:transformers.models.reformer.modeling_reformer:Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T07:48:47.2117642Z 2025-09-07T07:48:47.2117814Z loading model: 0it [00:01, ?it/s] 2025-09-07T07:48:47.2156609Z cpu eval hf_Reformer 2025-09-07T07:48:49.1761712Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:48:49.5004502Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:48:49.8221059Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:48:50.0968937Z class GraphModule(torch.nn.Module): 2025-09-07T07:48:50.0969295Z def forward(self, L_input_ids_: "i64[4, 2048][2048, 1]cpu"): 2025-09-07T07:48:50.0969623Z l_input_ids_ = L_input_ids_ 2025-09-07T07:48:50.0969863Z 2025-09-07T07:48:50.0970753Z # File: /opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py:194 in _get_least_common_mult_chunk_len, code: return np.lcm(config.lsh_attn_chunk_length, config.local_attn_chunk_length) 2025-09-07T07:48:50.0971867Z least_common_mult_chunk_length: "i64[][]cpu" = torch__dynamo_utils_wrapped_lcm(64, 64) 2025-09-07T07:48:50.0972259Z 2025-09-07T07:48:50.0972982Z # File: /opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py:2099 in forward, code: input_shape[-1] % least_common_mult_chunk_length != 0 2025-09-07T07:48:50.0974087Z wrapped_mod: "i64[][]cpu" = torch__dynamo_utils_wrapped_mod(2048, least_common_mult_chunk_length); least_common_mult_chunk_length = None 2025-09-07T07:48:50.0974843Z wrapped_ne: "b8[][]cpu" = torch__dynamo_utils_wrapped_ne(wrapped_mod, 0); wrapped_mod = wrapped_ne = None 2025-09-07T07:48:50.0975290Z 2025-09-07T07:48:50.0975480Z class GraphModule(torch.nn.Module): 2025-09-07T07:48:50.0975797Z def forward(self, L_input_ids_: "i64[4, 2048][2048, 1]cpu"): 2025-09-07T07:48:50.0976118Z l_input_ids_ = L_input_ids_ 2025-09-07T07:48:50.0976350Z 2025-09-07T07:48:50.0977213Z # File: /opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py:194 in _get_least_common_mult_chunk_len, code: return np.lcm(config.lsh_attn_chunk_length, config.local_attn_chunk_length) 2025-09-07T07:48:50.0978303Z least_common_mult_chunk_length: "i64[][]cpu" = torch__dynamo_utils_wrapped_lcm(64, 64) 2025-09-07T07:48:50.0978690Z 2025-09-07T07:48:50.0979409Z # File: /opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py:2099 in forward, code: input_shape[-1] % least_common_mult_chunk_length != 0 2025-09-07T07:48:50.0980498Z wrapped_mod: "i64[][]cpu" = torch__dynamo_utils_wrapped_mod(2048, least_common_mult_chunk_length); least_common_mult_chunk_length = None 2025-09-07T07:48:50.0981281Z wrapped_ne: "b8[][]cpu" = torch__dynamo_utils_wrapped_ne(wrapped_mod, 0); wrapped_mod = wrapped_ne = None 2025-09-07T07:48:50.0981720Z 2025-09-07T07:48:50.0981883Z ERROR:common: 2025-09-07T07:48:50.0982078Z Traceback (most recent call last): 2025-09-07T07:48:50.0982532Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 2320, in check_accuracy 2025-09-07T07:48:50.0983013Z optimized_model_iter_fn = optimize_ctx( 2025-09-07T07:48:50.0983458Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 1523, in export 2025-09-07T07:48:50.0983891Z ep = torch.export.export( 2025-09-07T07:48:50.0984629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/export/__init__.py", line 311, in export 2025-09-07T07:48:50.0985130Z raise e 2025-09-07T07:48:50.0985543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/export/__init__.py", line 277, in export 2025-09-07T07:48:50.0986095Z return _export( 2025-09-07T07:48:50.0986525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T07:48:50.0986995Z raise e 2025-09-07T07:48:50.0987398Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T07:48:50.0987878Z ep = fn(*args, **kwargs) 2025-09-07T07:48:50.0988369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T07:48:50.0988991Z return fn(*args, **kwargs) 2025-09-07T07:48:50.0989452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/export/_trace.py", line 2255, in _export 2025-09-07T07:48:50.0989936Z ep = _export_for_training( 2025-09-07T07:48:50.0990393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T07:48:50.0990861Z raise e 2025-09-07T07:48:50.0991265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T07:48:50.0991745Z ep = fn(*args, **kwargs) 2025-09-07T07:48:50.0992236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T07:48:50.0992756Z return fn(*args, **kwargs) 2025-09-07T07:48:50.0993268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/export/_trace.py", line 2071, in _export_for_training 2025-09-07T07:48:50.0993813Z export_artifact = export_func( 2025-09-07T07:48:50.0994318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/export/_trace.py", line 1415, in _strict_export 2025-09-07T07:48:50.0994842Z gm_torch_level = _export_to_torch_ir( 2025-09-07T07:48:50.0995380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/export/_trace.py", line 812, in _export_to_torch_ir 2025-09-07T07:48:50.0995930Z gm_torch_level, _ = torch._dynamo.export( 2025-09-07T07:48:50.0996443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 2002, in inner 2025-09-07T07:48:50.0996953Z result_traced = opt_f(*args, **kwargs) 2025-09-07T07:48:50.0997458Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 414, in __call__ 2025-09-07T07:48:50.0997975Z return super().__call__(*args, **kwargs) 2025-09-07T07:48:50.0998535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl 2025-09-07T07:48:50.0999108Z return self._call_impl(*args, **kwargs) 2025-09-07T07:48:50.0999628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T07:48:50.1000159Z return forward_call(*args, **kwargs) 2025-09-07T07:48:50.1000695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 841, in compile_wrapper 2025-09-07T07:48:50.1001322Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-09-07T07:48:50.1001749Z torch._dynamo.exc.Unsupported: Data-dependent branching 2025-09-07T07:48:50.1002377Z Explanation: Detected data-dependent branching (e.g. `if my_tensor.sum() > 0:`). Dynamo does not support tracing dynamic control flow. 2025-09-07T07:48:50.1003268Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-09-07T07:48:50.1003915Z Hint: Use `torch.cond` to express dynamic control flow. 2025-09-07T07:48:50.1004151Z 2025-09-07T07:48:50.1004356Z Developer debug context: attempted to jump with NumpyNdarrayVariable() 2025-09-07T07:48:50.1004653Z 2025-09-07T07:48:50.1005185Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0170.html 2025-09-07T07:48:50.1005686Z 2025-09-07T07:48:50.1005800Z from user code: 2025-09-07T07:48:50.1006348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 2481, in forward 2025-09-07T07:48:50.1006976Z reformer_outputs = self.reformer( 2025-09-07T07:48:50.1007494Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T07:48:50.1008025Z return forward_call(*args, **kwargs) 2025-09-07T07:48:50.1008634Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 2099, in forward 2025-09-07T07:48:50.1009337Z input_shape[-1] % least_common_mult_chunk_length != 0 2025-09-07T07:48:50.1009571Z 2025-09-07T07:48:50.1010100Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-09-07T07:48:50.1010725Z 2025-09-07T07:48:50.1010919Z TorchDynamo optimized model failed to run because of following error 2025-09-07T07:48:50.1257036Z fail_to_run 2025-09-07T07:48:50.1257548Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:48:51.2152940Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:48:51.2154232Z import pynvml # type: ignore[import] 2025-09-07T07:48:53.7743497Z 2025-09-07T07:49:03.8095059Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:49:03.8095359Z loading model: 0it [00:10, ?it/s] 2025-09-07T07:49:03.8312703Z cpu eval hf_Roberta_base 2025-09-07T07:49:09.6519594Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:49:10.6047749Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:49:11.5472341Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:49:17.0763626Z pass 2025-09-07T07:49:17.0764103Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:49:18.9618915Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:49:18.9620106Z import pynvml # type: ignore[import] 2025-09-07T07:49:21.5397742Z 2025-09-07T07:49:24.8995706Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:49:24.8996015Z loading model: 0it [00:03, ?it/s] 2025-09-07T07:49:24.9096622Z cpu eval hf_T5 2025-09-07T07:49:26.6339743Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:49:27.0598298Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:49:27.4829358Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:49:32.7035161Z pass 2025-09-07T07:49:32.7035567Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:49:34.0913959Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:49:34.0915534Z import pynvml # type: ignore[import] 2025-09-07T07:49:36.6536047Z 2025-09-07T07:49:48.6341972Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:49:48.6342587Z loading model: 0it [00:11, ?it/s] 2025-09-07T07:49:48.6594029Z cpu eval hf_T5_base 2025-09-07T07:50:23.4572158Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:50:32.2275179Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:50:40.9895824Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:50:59.4245949Z pass 2025-09-07T07:50:59.4246718Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:51:03.9123034Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:51:03.9124209Z import pynvml # type: ignore[import] 2025-09-07T07:51:06.5190942Z 2025-09-07T07:51:21.0556319Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:51:21.0556645Z loading model: 0it [00:14, ?it/s] 2025-09-07T07:51:21.1148743Z cpu eval hf_T5_large 2025-09-07T07:51:21.1441674Z pass_due_to_skip 2025-09-07T07:51:21.1444064Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:51:22.7174376Z accuracy pass_rate=60.00% 2025-09-07T07:51:22.7179606Z calls_captured gmean=0.00x mean=541.400x 2025-09-07T07:51:22.7183336Z unique_graphs gmean=0.00x mean=0.600x 2025-09-07T07:51:22.7187181Z graph_breaks gmean=0.00x mean=0.000x 2025-09-07T07:51:22.7190876Z unique_graph_breaks gmean=0.00x mean=0.000x 2025-09-07T07:51:22.7194622Z autograd_captures gmean=0.00x mean=0.000x 2025-09-07T07:51:22.7198281Z autograd_compiles gmean=0.00x mean=0.000x 2025-09-07T07:51:22.7201898Z cudagraph_skips gmean=0.00x mean=0.000x 2025-09-07T07:51:22.7203029Z compilation_latency mean=2.049 seconds 2025-09-07T07:51:23.1690591Z + taskset -c 0-62 python benchmarks/dynamo/torchbench.py --accuracy --no-translation-validation --inference --bfloat16 --export-aot-inductor --disable-cudagraphs --device cpu --total-partitions 15 --partition-id 7 --output /var/lib/jenkins/workspace/test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_aarch64_accuracy.csv 2025-09-07T07:51:23.5040619Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:51:23.5041781Z import pynvml # type: ignore[import] 2025-09-07T07:51:26.4159816Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:51:26.4161025Z import pynvml # type: ignore[import] 2025-09-07T07:51:28.9917626Z 2025-09-07T07:51:38.9906894Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:51:38.9907179Z loading model: 0it [00:09, ?it/s] 2025-09-07T07:51:39.0121266Z cpu eval hf_Roberta_base 2025-09-07T07:51:44.8164761Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:51:45.7712027Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:51:46.7101352Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:52:07.6234114Z pass 2025-09-07T07:52:07.6234971Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:52:10.2288631Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:52:10.2290155Z import pynvml # type: ignore[import] 2025-09-07T07:52:12.7895382Z 2025-09-07T07:52:16.1546859Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:52:16.1547164Z loading model: 0it [00:03, ?it/s] 2025-09-07T07:52:16.1647117Z cpu eval hf_T5 2025-09-07T07:52:17.8944261Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:52:18.3213672Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:52:18.7461210Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:52:36.8480235Z pass 2025-09-07T07:52:36.8480789Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:52:38.7775451Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:52:38.7776621Z import pynvml # type: ignore[import] 2025-09-07T07:52:41.3753878Z 2025-09-07T07:52:53.2283028Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:52:53.2283339Z loading model: 0it [00:11, ?it/s] 2025-09-07T07:52:53.2532867Z cpu eval hf_T5_base 2025-09-07T07:53:27.9306424Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:53:36.7676015Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:53:45.6011917Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:54:30.0119884Z pass 2025-09-07T07:54:30.0120379Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:54:35.4724783Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:54:35.4726047Z import pynvml # type: ignore[import] 2025-09-07T07:54:38.0678547Z 2025-09-07T07:54:52.6526393Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:54:52.6526720Z loading model: 0it [00:14, ?it/s] 2025-09-07T07:54:52.7179012Z cpu eval hf_T5_large 2025-09-07T07:54:52.7474023Z pass_due_to_skip 2025-09-07T07:54:52.7476655Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:54:54.3791050Z accuracy pass_rate=75.00% 2025-09-07T07:54:54.3796123Z calls_captured gmean=0.00x mean=0.000x 2025-09-07T07:54:54.3800006Z unique_graphs gmean=0.00x mean=0.000x 2025-09-07T07:54:54.3803824Z graph_breaks gmean=0.00x mean=0.000x 2025-09-07T07:54:54.3807516Z unique_graph_breaks gmean=0.00x mean=0.000x 2025-09-07T07:54:54.3811241Z autograd_captures gmean=0.00x mean=0.000x 2025-09-07T07:54:54.3814807Z autograd_compiles gmean=0.00x mean=0.000x 2025-09-07T07:54:54.3818436Z cudagraph_skips gmean=0.00x mean=0.000x 2025-09-07T07:54:54.3819393Z compilation_latency mean=0.000 seconds 2025-09-07T07:54:54.8286066Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *maxautotune-true* ]] 2025-09-07T07:54:54.8287538Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *cudagraphs_low_precision-true* ]] 2025-09-07T07:54:54.8288259Z + for target in "${targets[@]}" 2025-09-07T07:54:54.8288625Z + target_flag=('--performance') 2025-09-07T07:54:54.8288861Z + local target_flag 2025-09-07T07:54:54.8289079Z + [[ performance == \p\e\r\f\o\r\m\a\n\c\e ]] 2025-09-07T07:54:54.8289372Z + target_flag+=(--cold-start-latency) 2025-09-07T07:54:54.8289994Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *freezing-true* ]] 2025-09-07T07:54:54.8290967Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *default-true* ]] 2025-09-07T07:54:54.8292867Z + taskset -c 0-62 python benchmarks/dynamo/torchbench.py --performance --cold-start-latency --inference --bfloat16 --backend inductor --disable-cudagraphs --device cpu --total-partitions 15 --partition-id 7 --output /var/lib/jenkins/workspace/test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_aarch64_performance.csv 2025-09-07T07:54:55.1624374Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:54:55.1625553Z import pynvml # type: ignore[import] 2025-09-07T07:54:58.0656187Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:54:58.0657475Z import pynvml # type: ignore[import] 2025-09-07T07:55:00.6300021Z 2025-09-07T07:55:02.4451693Z loading model: 0it [00:00, ?it/s]Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T07:55:02.4453569Z WARNING:transformers.models.reformer.modeling_reformer:Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T07:55:02.5050192Z 2025-09-07T07:55:02.5050298Z loading model: 0it [00:01, ?it/s] 2025-09-07T07:55:02.5086229Z cpu eval hf_Reformer 2025-09-07T07:55:08.0739959Z W0907 07:55:08.073000 19022 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] Graph break from `Tensor.item()`, consider setting: 2025-09-07T07:55:08.0740930Z W0907 07:55:08.073000 19022 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] torch._dynamo.config.capture_scalar_outputs = True 2025-09-07T07:55:08.0741688Z W0907 07:55:08.073000 19022 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] or: 2025-09-07T07:55:08.0742395Z W0907 07:55:08.073000 19022 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1 2025-09-07T07:55:08.0743260Z W0907 07:55:08.073000 19022 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] to include these operations in the captured graph. 2025-09-07T07:55:08.0743968Z W0907 07:55:08.073000 19022 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T07:55:08.0744623Z W0907 07:55:08.073000 19022 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] Graph break: from user code at: 2025-09-07T07:55:08.0745892Z W0907 07:55:08.073000 19022 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 2132, in torch_dynamo_resume_in_forward_at_2104 2025-09-07T07:55:08.0747465Z W0907 07:55:08.073000 19022 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] embedding_output = self.embeddings( 2025-09-07T07:55:08.0748637Z W0907 07:55:08.073000 19022 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 369, in forward 2025-09-07T07:55:08.0749955Z W0907 07:55:08.073000 19022 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] position_embeddings = self.position_embeddings(position_ids) 2025-09-07T07:55:08.0751186Z W0907 07:55:08.073000 19022 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 293, in forward 2025-09-07T07:55:08.0752450Z W0907 07:55:08.073000 19022 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] max_position_id = position_ids.max().item() 2025-09-07T07:55:08.0753147Z W0907 07:55:08.073000 19022 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T07:55:08.0753695Z W0907 07:55:08.073000 19022 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T07:55:20.6049410Z 2025-09-07T07:55:20.7939324Z running benchmark: 0% 0/30 [00:00 2025-09-07T08:00:41.7112331Z torch.index_select(position_encodings[i], 0, position_ids[i]).unsqueeze(0) 2025-09-07T08:00:41.7112640Z 2025-09-07T08:00:42.3288036Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:42.3288841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 370, in torch_dynamo_resume_in_forward_at_369 2025-09-07T08:00:42.3289602Z embeddings = embeddings + position_embeddings 2025-09-07T08:00:42.3289817Z 2025-09-07T08:00:48.3344713Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3345112Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3345907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3346686Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3347348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3347958Z layer_outputs = layer( 2025-09-07T08:00:48.3348543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3349165Z attn_outputs = self.attention( 2025-09-07T08:00:48.3349768Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3350414Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3351051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:00:48.3351851Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:00:48.3352805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 807, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:00:48.3353647Z undo_sorted_bucket_idx.scatter_(-1, sorted_bucket_idx, indices) 2025-09-07T08:00:48.3353918Z 2025-09-07T08:00:48.3354056Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3354833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3355587Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3356230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3356836Z layer_outputs = layer( 2025-09-07T08:00:48.3357412Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3358371Z attn_outputs = self.attention( 2025-09-07T08:00:48.3358998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3359739Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3360380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:00:48.3361179Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:00:48.3362120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 807, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:00:48.3363062Z undo_sorted_bucket_idx.scatter_(-1, sorted_bucket_idx, indices) 2025-09-07T08:00:48.3363336Z 2025-09-07T08:00:48.3363473Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3364250Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3365005Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3365784Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3366390Z layer_outputs = layer( 2025-09-07T08:00:48.3366963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3367576Z attn_outputs = self.attention( 2025-09-07T08:00:48.3368176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3368822Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3369463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:00:48.3370264Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:00:48.3371211Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 807, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:00:48.3372047Z undo_sorted_bucket_idx.scatter_(-1, sorted_bucket_idx, indices) 2025-09-07T08:00:48.3372313Z 2025-09-07T08:00:48.3372670Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3373436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3374210Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3374845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3375460Z layer_outputs = layer( 2025-09-07T08:00:48.3384079Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3384822Z attn_outputs = self.attention( 2025-09-07T08:00:48.3385434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3386079Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3386713Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1198, in forward 2025-09-07T08:00:48.3387336Z query_vectors = self.query(hidden_states) 2025-09-07T08:00:48.3387545Z 2025-09-07T08:00:48.3387682Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3388723Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3389489Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3390129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3390814Z layer_outputs = layer( 2025-09-07T08:00:48.3391381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3391989Z attn_outputs = self.attention( 2025-09-07T08:00:48.3392583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3393306Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3393938Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1199, in forward 2025-09-07T08:00:48.3394554Z key_vectors = self.key(hidden_states) 2025-09-07T08:00:48.3394748Z 2025-09-07T08:00:48.3394846Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3395123Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3395440Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3396203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3396954Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3397590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3398193Z layer_outputs = layer( 2025-09-07T08:00:48.3398760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3399370Z attn_outputs = self.attention( 2025-09-07T08:00:48.3399969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3400604Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3401236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:00:48.3402012Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3402819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:00:48.3403569Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:00:48.3403882Z 2025-09-07T08:00:48.3403981Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3404260Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3404578Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3405346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3406098Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3406732Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3407337Z layer_outputs = layer( 2025-09-07T08:00:48.3407902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3408509Z attn_outputs = self.attention( 2025-09-07T08:00:48.3409105Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3409739Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3410494Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:00:48.3411279Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3412122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:00:48.3412758Z return torch.cat(slices, dim=3) 2025-09-07T08:00:48.3412930Z 2025-09-07T08:00:48.3413068Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3413831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3414623Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3415258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3415860Z layer_outputs = layer( 2025-09-07T08:00:48.3416421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3417030Z attn_outputs = self.attention( 2025-09-07T08:00:48.3417623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3418259Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3418890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:00:48.3419616Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:00:48.3419924Z 2025-09-07T08:00:48.3420058Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3420824Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3421569Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3422201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3422800Z layer_outputs = layer( 2025-09-07T08:00:48.3423362Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3423971Z attn_outputs = self.attention( 2025-09-07T08:00:48.3424563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3425202Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3425838Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:00:48.3426559Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:00:48.3426868Z 2025-09-07T08:00:48.3427005Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3427768Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3428517Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3429152Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3429749Z layer_outputs = layer( 2025-09-07T08:00:48.3430318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3430926Z attn_outputs = self.attention( 2025-09-07T08:00:48.3431639Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3432281Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3432976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1200, in forward 2025-09-07T08:00:48.3433603Z value_vectors = self.value(hidden_states) 2025-09-07T08:00:48.3433804Z 2025-09-07T08:00:48.3433902Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3434181Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3434493Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3435255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3436060Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3436697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3437300Z layer_outputs = layer( 2025-09-07T08:00:48.3437862Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3438473Z attn_outputs = self.attention( 2025-09-07T08:00:48.3439065Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3439697Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3440330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T08:00:48.3441128Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3441956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:00:48.3442711Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:00:48.3443019Z 2025-09-07T08:00:48.3443119Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3443399Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3443709Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3444475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3445221Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3445855Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3446462Z layer_outputs = layer( 2025-09-07T08:00:48.3447027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3447635Z attn_outputs = self.attention( 2025-09-07T08:00:48.3448230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3448866Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3449499Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:00:48.3450167Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:00:48.3450422Z 2025-09-07T08:00:48.3450556Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3451321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3452066Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3452816Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3453426Z layer_outputs = layer( 2025-09-07T08:00:48.3454034Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3454640Z attn_outputs = self.attention( 2025-09-07T08:00:48.3455232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3455861Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3456491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:00:48.3457197Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:00:48.3457444Z 2025-09-07T08:00:48.3457582Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3458354Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3459097Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3459728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3460329Z layer_outputs = layer( 2025-09-07T08:00:48.3460896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3461503Z attn_outputs = self.attention( 2025-09-07T08:00:48.3462093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3462725Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3463360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T08:00:48.3464182Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:00:48.3465071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:00:48.3465947Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:00:48.3466244Z 2025-09-07T08:00:48.3466379Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3467143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3467892Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3468529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3469124Z layer_outputs = layer( 2025-09-07T08:00:48.3469689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3470301Z attn_outputs = self.attention( 2025-09-07T08:00:48.3470896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T08:00:48.3471605Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T08:00:48.3472314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T08:00:48.3472941Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3473142Z 2025-09-07T08:00:48.3473274Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3474219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3474978Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3475669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3476270Z layer_outputs = layer( 2025-09-07T08:00:48.3476839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.3477519Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.3478205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.3478898Z return apply_chunking_to_forward( 2025-09-07T08:00:48.3479489Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.3480079Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.3480704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:00:48.3481356Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3481970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T08:00:48.3482594Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3482792Z 2025-09-07T08:00:48.3482926Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3483695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3484441Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3485071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3485673Z layer_outputs = layer( 2025-09-07T08:00:48.3486245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.3486927Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.3487606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.3488220Z return apply_chunking_to_forward( 2025-09-07T08:00:48.3488803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.3489397Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.3490025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:00:48.3490675Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3491296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T08:00:48.3491925Z hidden_states = self.act_fn(hidden_states) 2025-09-07T08:00:48.3492123Z 2025-09-07T08:00:48.3492263Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3493043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3493791Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3494429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3495030Z layer_outputs = layer( 2025-09-07T08:00:48.3495716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.3496415Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.3497137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.3497755Z return apply_chunking_to_forward( 2025-09-07T08:00:48.3498336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.3498920Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.3499544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T08:00:48.3500222Z return self.output(hidden_states) 2025-09-07T08:00:48.3500829Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T08:00:48.3501454Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3501652Z 2025-09-07T08:00:48.3501789Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3502551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3503298Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3503930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3504533Z layer_outputs = layer( 2025-09-07T08:00:48.3505101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3505711Z attn_outputs = self.attention( 2025-09-07T08:00:48.3506307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3506942Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3507570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 540, in forward 2025-09-07T08:00:48.3508206Z query_key_vectors = self.query_key(hidden_states) 2025-09-07T08:00:48.3508429Z 2025-09-07T08:00:48.3508562Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3509322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3510076Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3510705Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3511304Z layer_outputs = layer( 2025-09-07T08:00:48.3511873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3512489Z attn_outputs = self.attention( 2025-09-07T08:00:48.3513080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3513712Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3514346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:00:48.3515063Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:00:48.3515806Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T08:00:48.3516558Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T08:00:48.3517019Z 2025-09-07T08:00:48.3517164Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3517940Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3518730Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3519363Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3519965Z layer_outputs = layer( 2025-09-07T08:00:48.3520531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3521186Z attn_outputs = self.attention( 2025-09-07T08:00:48.3521779Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3522415Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3523048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:00:48.3523765Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:00:48.3524511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T08:00:48.3525249Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T08:00:48.3525544Z 2025-09-07T08:00:48.3525683Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3526455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3527205Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3527841Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3528442Z layer_outputs = layer( 2025-09-07T08:00:48.3529015Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3529626Z attn_outputs = self.attention( 2025-09-07T08:00:48.3530217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3530852Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3531482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:00:48.3532203Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:00:48.3532953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T08:00:48.3533609Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T08:00:48.3533830Z 2025-09-07T08:00:48.3533966Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3534728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3535476Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3536108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3536708Z layer_outputs = layer( 2025-09-07T08:00:48.3537280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3537891Z attn_outputs = self.attention( 2025-09-07T08:00:48.3538608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3539260Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3539929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:00:48.3540717Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:00:48.3541650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:00:48.3542450Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:00:48.3543181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T08:00:48.3543950Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T08:00:48.3544267Z 2025-09-07T08:00:48.3544369Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3544690Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3545459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3546204Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3546833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3547435Z layer_outputs = layer( 2025-09-07T08:00:48.3548008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3548620Z attn_outputs = self.attention( 2025-09-07T08:00:48.3549215Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3549851Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3550484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:00:48.3551273Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:00:48.3552209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:00:48.3553004Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:00:48.3553683Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T08:00:48.3554347Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T08:00:48.3554556Z 2025-09-07T08:00:48.3554658Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3554939Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3555252Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3556020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3556769Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3557403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3558008Z layer_outputs = layer( 2025-09-07T08:00:48.3558577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3559187Z attn_outputs = self.attention( 2025-09-07T08:00:48.3559902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3560555Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3561226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3561872Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3562525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:00:48.3563344Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3564245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:00:48.3564994Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:00:48.3565374Z 2025-09-07T08:00:48.3565478Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3565758Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3566078Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3566844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3567590Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3568221Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3568820Z layer_outputs = layer( 2025-09-07T08:00:48.3569389Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3569999Z attn_outputs = self.attention( 2025-09-07T08:00:48.3570592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3571228Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3571863Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3572509Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3573151Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:00:48.3573961Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3574812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:00:48.3575446Z return torch.cat(slices, dim=3) 2025-09-07T08:00:48.3575617Z 2025-09-07T08:00:48.3575718Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3576035Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3576799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3577549Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3578180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3578779Z layer_outputs = layer( 2025-09-07T08:00:48.3579352Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3579962Z attn_outputs = self.attention( 2025-09-07T08:00:48.3580558Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3581389Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3582054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 541, in forward 2025-09-07T08:00:48.3582733Z value_vectors = self.value(hidden_states) 2025-09-07T08:00:48.3582931Z 2025-09-07T08:00:48.3583064Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3583826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3584569Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3585203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3585869Z layer_outputs = layer( 2025-09-07T08:00:48.3586436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3587045Z attn_outputs = self.attention( 2025-09-07T08:00:48.3587635Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3588271Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3588898Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3589549Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3590195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:00:48.3590973Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3591782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:00:48.3592532Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:00:48.3592842Z 2025-09-07T08:00:48.3592940Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3593258Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3594023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3594767Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3595397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3596006Z layer_outputs = layer( 2025-09-07T08:00:48.3596575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3597184Z attn_outputs = self.attention( 2025-09-07T08:00:48.3597780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3598417Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3599049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3599694Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3600335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:00:48.3601107Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3601910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:00:48.3602545Z return torch.cat(slices, dim=3) 2025-09-07T08:00:48.3602861Z 2025-09-07T08:00:48.3603003Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3603766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3604551Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3605182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3605783Z layer_outputs = layer( 2025-09-07T08:00:48.3606353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3607000Z attn_outputs = self.attention( 2025-09-07T08:00:48.3607595Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3608237Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3608866Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T08:00:48.3609685Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T08:00:48.3610564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T08:00:48.3611263Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T08:00:48.3611476Z 2025-09-07T08:00:48.3611612Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3612382Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3613132Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3613769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3614378Z layer_outputs = layer( 2025-09-07T08:00:48.3614941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3615548Z attn_outputs = self.attention( 2025-09-07T08:00:48.3616141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3616778Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3617408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3618064Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3618719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T08:00:48.3619445Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:00:48.3619751Z 2025-09-07T08:00:48.3619893Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3620657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3621413Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3622047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3622653Z layer_outputs = layer( 2025-09-07T08:00:48.3623219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3623822Z attn_outputs = self.attention( 2025-09-07T08:00:48.3624534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3625177Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3625846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3626496Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3627137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T08:00:48.3627930Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3628795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:00:48.3629553Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:00:48.3629860Z 2025-09-07T08:00:48.3629962Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3630279Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3631048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3631793Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3632427Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3633028Z layer_outputs = layer( 2025-09-07T08:00:48.3633588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3634202Z attn_outputs = self.attention( 2025-09-07T08:00:48.3634800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3635441Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3636076Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3636721Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3637364Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:00:48.3638031Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:00:48.3638279Z 2025-09-07T08:00:48.3638419Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3639185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3639935Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3640575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3641185Z layer_outputs = layer( 2025-09-07T08:00:48.3641753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3642365Z attn_outputs = self.attention( 2025-09-07T08:00:48.3642960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3643597Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3644228Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3644879Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3645645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:00:48.3646336Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:00:48.3646622Z 2025-09-07T08:00:48.3646719Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3647038Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3647804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3648551Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3649183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3649817Z layer_outputs = layer( 2025-09-07T08:00:48.3650381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3650993Z attn_outputs = self.attention( 2025-09-07T08:00:48.3651583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3652220Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3652850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T08:00:48.3653664Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:00:48.3654552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:00:48.3655332Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:00:48.3655627Z 2025-09-07T08:00:48.3655761Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3656534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3657288Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3657918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3658521Z layer_outputs = layer( 2025-09-07T08:00:48.3659088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3659695Z attn_outputs = self.attention( 2025-09-07T08:00:48.3660289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T08:00:48.3661001Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T08:00:48.3661717Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T08:00:48.3662342Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3662543Z 2025-09-07T08:00:48.3662677Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3663441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3664192Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3664822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3665524Z layer_outputs = layer( 2025-09-07T08:00:48.3666098Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.3666947Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.3667643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.3668303Z return apply_chunking_to_forward( 2025-09-07T08:00:48.3668887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.3669478Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.3670103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:00:48.3670753Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3671461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T08:00:48.3672087Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3672287Z 2025-09-07T08:00:48.3672425Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3673186Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3673936Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3674567Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3675168Z layer_outputs = layer( 2025-09-07T08:00:48.3675738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.3676418Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.3677101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.3677714Z return apply_chunking_to_forward( 2025-09-07T08:00:48.3678296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.3678886Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.3679511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:00:48.3680157Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3680777Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T08:00:48.3681399Z hidden_states = self.act_fn(hidden_states) 2025-09-07T08:00:48.3681601Z 2025-09-07T08:00:48.3681740Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3682503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3683249Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3683880Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3684486Z layer_outputs = layer( 2025-09-07T08:00:48.3685058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.3685743Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.3686426Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.3687044Z return apply_chunking_to_forward( 2025-09-07T08:00:48.3687626Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.3688216Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.3688949Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T08:00:48.3689685Z return self.output(hidden_states) 2025-09-07T08:00:48.3690292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T08:00:48.3690914Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3691110Z 2025-09-07T08:00:48.3691245Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3692001Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3692805Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3693440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3694045Z layer_outputs = layer( 2025-09-07T08:00:48.3694612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3695222Z attn_outputs = self.attention( 2025-09-07T08:00:48.3695812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3696450Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3697083Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1198, in forward 2025-09-07T08:00:48.3697708Z query_vectors = self.query(hidden_states) 2025-09-07T08:00:48.3697906Z 2025-09-07T08:00:48.3698040Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3698806Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3699555Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3700188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3700795Z layer_outputs = layer( 2025-09-07T08:00:48.3701358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3701965Z attn_outputs = self.attention( 2025-09-07T08:00:48.3702559Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3703196Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3703830Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1199, in forward 2025-09-07T08:00:48.3704443Z key_vectors = self.key(hidden_states) 2025-09-07T08:00:48.3704632Z 2025-09-07T08:00:48.3704730Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3705010Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3705328Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3706085Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3706833Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3707466Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3708070Z layer_outputs = layer( 2025-09-07T08:00:48.3708639Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3709247Z attn_outputs = self.attention( 2025-09-07T08:00:48.3709955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3710643Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3711279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:00:48.3712057Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3712861Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:00:48.3713614Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:00:48.3713964Z 2025-09-07T08:00:48.3714063Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3714343Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3714662Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3715425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3716176Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3716806Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3717409Z layer_outputs = layer( 2025-09-07T08:00:48.3717979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3718585Z attn_outputs = self.attention( 2025-09-07T08:00:48.3719182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3719813Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3720447Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:00:48.3721222Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3722030Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:00:48.3722670Z return torch.cat(slices, dim=3) 2025-09-07T08:00:48.3722846Z 2025-09-07T08:00:48.3722980Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3723745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3724500Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3725127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3725734Z layer_outputs = layer( 2025-09-07T08:00:48.3726301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3726915Z attn_outputs = self.attention( 2025-09-07T08:00:48.3727505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3728141Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3728776Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:00:48.3729503Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:00:48.3729810Z 2025-09-07T08:00:48.3729948Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3730825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3731583Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3732255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3732860Z layer_outputs = layer( 2025-09-07T08:00:48.3733429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3734036Z attn_outputs = self.attention( 2025-09-07T08:00:48.3734628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3735301Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3735937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:00:48.3736664Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:00:48.3736969Z 2025-09-07T08:00:48.3737105Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3737868Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3738614Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3739247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3739851Z layer_outputs = layer( 2025-09-07T08:00:48.3740414Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3741019Z attn_outputs = self.attention( 2025-09-07T08:00:48.3741613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3742251Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3742886Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1200, in forward 2025-09-07T08:00:48.3743507Z value_vectors = self.value(hidden_states) 2025-09-07T08:00:48.3743708Z 2025-09-07T08:00:48.3743804Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3744085Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3744399Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3745165Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3745916Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3746553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3747154Z layer_outputs = layer( 2025-09-07T08:00:48.3747720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3748328Z attn_outputs = self.attention( 2025-09-07T08:00:48.3748925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3749565Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3750201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T08:00:48.3751000Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3751933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:00:48.3752697Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:00:48.3753065Z 2025-09-07T08:00:48.3753163Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3753443Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3753758Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3754525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3755272Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3755905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3756551Z layer_outputs = layer( 2025-09-07T08:00:48.3757128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3757737Z attn_outputs = self.attention( 2025-09-07T08:00:48.3758334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3758973Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3759607Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:00:48.3760273Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:00:48.3760524Z 2025-09-07T08:00:48.3760658Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3761423Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3762175Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3762815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3763419Z layer_outputs = layer( 2025-09-07T08:00:48.3763986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3764594Z attn_outputs = self.attention( 2025-09-07T08:00:48.3765189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3765943Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3766578Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:00:48.3767252Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:00:48.3767502Z 2025-09-07T08:00:48.3767636Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3768407Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3769157Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3769785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3770388Z layer_outputs = layer( 2025-09-07T08:00:48.3770956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3771564Z attn_outputs = self.attention( 2025-09-07T08:00:48.3772153Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3772786Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3773599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T08:00:48.3774432Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:00:48.3775375Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:00:48.3776153Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:00:48.3776443Z 2025-09-07T08:00:48.3776580Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3777345Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3778163Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3778804Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3779410Z layer_outputs = layer( 2025-09-07T08:00:48.3779975Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3780586Z attn_outputs = self.attention( 2025-09-07T08:00:48.3781181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T08:00:48.3781893Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T08:00:48.3782606Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T08:00:48.3783231Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3783433Z 2025-09-07T08:00:48.3783568Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3784336Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3785083Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3785722Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3786323Z layer_outputs = layer( 2025-09-07T08:00:48.3786890Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.3787578Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.3788267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.3788886Z return apply_chunking_to_forward( 2025-09-07T08:00:48.3789471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.3790065Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.3790691Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:00:48.3791341Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3791957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T08:00:48.3792583Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3792784Z 2025-09-07T08:00:48.3792917Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3793680Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3794431Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3795234Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3795863Z layer_outputs = layer( 2025-09-07T08:00:48.3796483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.3797163Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.3797845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.3798456Z return apply_chunking_to_forward( 2025-09-07T08:00:48.3799037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.3799667Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.3800296Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:00:48.3800947Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3801564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T08:00:48.3802189Z hidden_states = self.act_fn(hidden_states) 2025-09-07T08:00:48.3802389Z 2025-09-07T08:00:48.3802522Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3803287Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3804029Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3804658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3805258Z layer_outputs = layer( 2025-09-07T08:00:48.3805826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.3806512Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.3807201Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.3807812Z return apply_chunking_to_forward( 2025-09-07T08:00:48.3808396Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.3808983Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.3809608Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T08:00:48.3810250Z return self.output(hidden_states) 2025-09-07T08:00:48.3810852Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T08:00:48.3811475Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3811676Z 2025-09-07T08:00:48.3811810Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3812577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3813325Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3813956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3814559Z layer_outputs = layer( 2025-09-07T08:00:48.3815132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3815740Z attn_outputs = self.attention( 2025-09-07T08:00:48.3816445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3817093Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3817766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 540, in forward 2025-09-07T08:00:48.3818405Z query_key_vectors = self.query_key(hidden_states) 2025-09-07T08:00:48.3818622Z 2025-09-07T08:00:48.3818761Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3819518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3820310Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3820941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3821541Z layer_outputs = layer( 2025-09-07T08:00:48.3822111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3822719Z attn_outputs = self.attention( 2025-09-07T08:00:48.3823314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3823953Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3824591Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:00:48.3825306Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:00:48.3826054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T08:00:48.3826813Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T08:00:48.3827129Z 2025-09-07T08:00:48.3827269Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3828037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3828786Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3829416Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3830019Z layer_outputs = layer( 2025-09-07T08:00:48.3830587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3831196Z attn_outputs = self.attention( 2025-09-07T08:00:48.3831791Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3832425Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3833055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:00:48.3833774Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:00:48.3834526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T08:00:48.3835264Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T08:00:48.3835557Z 2025-09-07T08:00:48.3835694Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3836462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3837214Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3837980Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3838594Z layer_outputs = layer( 2025-09-07T08:00:48.3839205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3839817Z attn_outputs = self.attention( 2025-09-07T08:00:48.3847793Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3848532Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3849166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:00:48.3850013Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:00:48.3850760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T08:00:48.3851421Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T08:00:48.3851635Z 2025-09-07T08:00:48.3851774Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3852538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3853282Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3853910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3854512Z layer_outputs = layer( 2025-09-07T08:00:48.3855080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3855686Z attn_outputs = self.attention( 2025-09-07T08:00:48.3856286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3856917Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3857544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:00:48.3858330Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:00:48.3859266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:00:48.3860055Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:00:48.3860729Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T08:00:48.3861493Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T08:00:48.3861806Z 2025-09-07T08:00:48.3862059Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3862380Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3863150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3863891Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3864519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3865117Z layer_outputs = layer( 2025-09-07T08:00:48.3865814Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3866430Z attn_outputs = self.attention( 2025-09-07T08:00:48.3867205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3867849Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3868543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:00:48.3869329Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:00:48.3870262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:00:48.3871051Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:00:48.3871795Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T08:00:48.3872458Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T08:00:48.3872664Z 2025-09-07T08:00:48.3872771Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3873046Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3873357Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3874124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3874869Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3875498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3876102Z layer_outputs = layer( 2025-09-07T08:00:48.3876666Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3877276Z attn_outputs = self.attention( 2025-09-07T08:00:48.3877873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3878511Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3879139Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3879789Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3880435Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:00:48.3881250Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3882093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:00:48.3882848Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:00:48.3883154Z 2025-09-07T08:00:48.3883257Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3883539Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3883850Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3884621Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3885366Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3886005Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3886608Z layer_outputs = layer( 2025-09-07T08:00:48.3887175Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3887786Z attn_outputs = self.attention( 2025-09-07T08:00:48.3890959Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3891692Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3892433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3893084Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3893724Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:00:48.3894536Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3895383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:00:48.3896069Z return torch.cat(slices, dim=3) 2025-09-07T08:00:48.3896241Z 2025-09-07T08:00:48.3896341Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3896676Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3897443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3898187Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3898818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3899421Z layer_outputs = layer( 2025-09-07T08:00:48.3899988Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3900598Z attn_outputs = self.attention( 2025-09-07T08:00:48.3901184Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3901826Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3902455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 541, in forward 2025-09-07T08:00:48.3903075Z value_vectors = self.value(hidden_states) 2025-09-07T08:00:48.3903270Z 2025-09-07T08:00:48.3903405Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3904169Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3904915Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3905546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3906149Z layer_outputs = layer( 2025-09-07T08:00:48.3906715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3907323Z attn_outputs = self.attention( 2025-09-07T08:00:48.3907913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3908556Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3909182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3909830Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3910467Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:00:48.3911242Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3912048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:00:48.3912965Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:00:48.3913283Z 2025-09-07T08:00:48.3913425Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3913737Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3914502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3915246Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3915878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3916559Z layer_outputs = layer( 2025-09-07T08:00:48.3917124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3917734Z attn_outputs = self.attention( 2025-09-07T08:00:48.3918332Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3918974Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3919599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3920246Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3920882Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:00:48.3921653Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3922461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:00:48.3923094Z return torch.cat(slices, dim=3) 2025-09-07T08:00:48.3923275Z 2025-09-07T08:00:48.3923413Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3924175Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3924923Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3925551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3926149Z layer_outputs = layer( 2025-09-07T08:00:48.3926711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3927323Z attn_outputs = self.attention( 2025-09-07T08:00:48.3927915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3928550Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3929182Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T08:00:48.3929994Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T08:00:48.3930864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T08:00:48.3931556Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T08:00:48.3931767Z 2025-09-07T08:00:48.3931903Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3932666Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3933412Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3934185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3934796Z layer_outputs = layer( 2025-09-07T08:00:48.3935401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3936006Z attn_outputs = self.attention( 2025-09-07T08:00:48.3936603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3937240Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3937867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3938556Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3939200Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T08:00:48.3939926Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:00:48.3940237Z 2025-09-07T08:00:48.3940371Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3941134Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3941882Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3942515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3943121Z layer_outputs = layer( 2025-09-07T08:00:48.3943689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3944296Z attn_outputs = self.attention( 2025-09-07T08:00:48.3944897Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3945530Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3946164Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3946809Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3947454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T08:00:48.3948246Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.3949071Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:00:48.3949824Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:00:48.3950133Z 2025-09-07T08:00:48.3950234Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3950551Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3951322Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3952062Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3952692Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3953297Z layer_outputs = layer( 2025-09-07T08:00:48.3953864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3954475Z attn_outputs = self.attention( 2025-09-07T08:00:48.3955144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3955837Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3956475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3957156Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3957793Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:00:48.3958457Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:00:48.3958706Z 2025-09-07T08:00:48.3958840Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3959632Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3960379Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3961012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3961618Z layer_outputs = layer( 2025-09-07T08:00:48.3962185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3962791Z attn_outputs = self.attention( 2025-09-07T08:00:48.3963383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3964014Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3964645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.3965380Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.3966029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:00:48.3966702Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:00:48.3966951Z 2025-09-07T08:00:48.3967049Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.3967364Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3968124Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3968869Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3969498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3970098Z layer_outputs = layer( 2025-09-07T08:00:48.3970665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3971273Z attn_outputs = self.attention( 2025-09-07T08:00:48.3971867Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.3972500Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.3973127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T08:00:48.3973938Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:00:48.3974818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:00:48.3975600Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:00:48.3975891Z 2025-09-07T08:00:48.3976029Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3976954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3977758Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3978390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3978989Z layer_outputs = layer( 2025-09-07T08:00:48.3979556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.3980159Z attn_outputs = self.attention( 2025-09-07T08:00:48.3980750Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T08:00:48.3981541Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T08:00:48.3982258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T08:00:48.3982879Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3983081Z 2025-09-07T08:00:48.3983215Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3983974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3984717Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3985342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3985940Z layer_outputs = layer( 2025-09-07T08:00:48.3986503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1601, in forward 2025-09-07T08:00:48.3987133Z attn_output = prev_attn_output + attn_output 2025-09-07T08:00:48.3987340Z 2025-09-07T08:00:48.3987477Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3988235Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3988977Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3989605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3990206Z layer_outputs = layer( 2025-09-07T08:00:48.3990773Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.3991460Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.3992138Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.3992753Z return apply_chunking_to_forward( 2025-09-07T08:00:48.3993334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.3993924Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.3994546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:00:48.3995191Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3995810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T08:00:48.3996434Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.3996634Z 2025-09-07T08:00:48.3996770Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.3997592Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.3998389Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.3999022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.3999667Z layer_outputs = layer( 2025-09-07T08:00:48.4000234Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.4000913Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.4001590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.4002231Z return apply_chunking_to_forward( 2025-09-07T08:00:48.4002811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.4003403Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.4004026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:00:48.4004675Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.4005292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T08:00:48.4005913Z hidden_states = self.act_fn(hidden_states) 2025-09-07T08:00:48.4006110Z 2025-09-07T08:00:48.4006246Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4007003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4007748Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4008377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4008974Z layer_outputs = layer( 2025-09-07T08:00:48.4009540Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.4010221Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.4010905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.4011515Z return apply_chunking_to_forward( 2025-09-07T08:00:48.4012094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.4012683Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.4013305Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T08:00:48.4013948Z return self.output(hidden_states) 2025-09-07T08:00:48.4014551Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T08:00:48.4015175Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.4015180Z 2025-09-07T08:00:48.4015313Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4015850Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4015963Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4016388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4016475Z layer_outputs = layer( 2025-09-07T08:00:48.4016970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.4017182Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.4017188Z 2025-09-07T08:00:48.4017359Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4017896Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4018009Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4018433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4018517Z layer_outputs = layer( 2025-09-07T08:00:48.4018977Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4019071Z attn_outputs = self.attention( 2025-09-07T08:00:48.4019498Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4019618Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4020041Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1198, in forward 2025-09-07T08:00:48.4020143Z query_vectors = self.query(hidden_states) 2025-09-07T08:00:48.4020148Z 2025-09-07T08:00:48.4020284Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4020817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4020935Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4021361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4021440Z layer_outputs = layer( 2025-09-07T08:00:48.4021869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4021961Z attn_outputs = self.attention( 2025-09-07T08:00:48.4022387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4022502Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4022923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1199, in forward 2025-09-07T08:00:48.4023018Z key_vectors = self.key(hidden_states) 2025-09-07T08:00:48.4023022Z 2025-09-07T08:00:48.4023121Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4023221Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4023355Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4023898Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4024011Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4024436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4024517Z layer_outputs = layer( 2025-09-07T08:00:48.4024939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4025032Z attn_outputs = self.attention( 2025-09-07T08:00:48.4025455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4025577Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4026108Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:00:48.4026422Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.4026906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:00:48.4027110Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:00:48.4027116Z 2025-09-07T08:00:48.4027215Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4027311Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4027443Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4027984Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4028123Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4028555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4028635Z layer_outputs = layer( 2025-09-07T08:00:48.4029063Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4029152Z attn_outputs = self.attention( 2025-09-07T08:00:48.4029574Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4029692Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4030115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:00:48.4030377Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.4030827Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:00:48.4030920Z return torch.cat(slices, dim=3) 2025-09-07T08:00:48.4030927Z 2025-09-07T08:00:48.4031062Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4031595Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4031711Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4032138Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4032219Z layer_outputs = layer( 2025-09-07T08:00:48.4032642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4032731Z attn_outputs = self.attention( 2025-09-07T08:00:48.4033156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4033271Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4033696Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:00:48.4033903Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:00:48.4033908Z 2025-09-07T08:00:48.4034047Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4034584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4034700Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4035125Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4035261Z layer_outputs = layer( 2025-09-07T08:00:48.4035731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4035846Z attn_outputs = self.attention( 2025-09-07T08:00:48.4036275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4036390Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4036811Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:00:48.4037017Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:00:48.4037047Z 2025-09-07T08:00:48.4037186Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4037727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4037843Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4038267Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4038348Z layer_outputs = layer( 2025-09-07T08:00:48.4038770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4038861Z attn_outputs = self.attention( 2025-09-07T08:00:48.4039281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4039399Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4039821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1200, in forward 2025-09-07T08:00:48.4039926Z value_vectors = self.value(hidden_states) 2025-09-07T08:00:48.4039932Z 2025-09-07T08:00:48.4040034Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4040131Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4040267Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4040799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4040912Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4041338Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4041418Z layer_outputs = layer( 2025-09-07T08:00:48.4041844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4041932Z attn_outputs = self.attention( 2025-09-07T08:00:48.4042360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4042479Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4042900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T08:00:48.4043178Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.4043631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:00:48.4043837Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:00:48.4043845Z 2025-09-07T08:00:48.4043943Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4044043Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4044234Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4044835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4044984Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4045408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4045491Z layer_outputs = layer( 2025-09-07T08:00:48.4045913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4046001Z attn_outputs = self.attention( 2025-09-07T08:00:48.4046460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4046574Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4047000Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:00:48.4047153Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:00:48.4047158Z 2025-09-07T08:00:48.4047295Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4047828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4047940Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4048368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4048449Z layer_outputs = layer( 2025-09-07T08:00:48.4048874Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4048965Z attn_outputs = self.attention( 2025-09-07T08:00:48.4049388Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4049509Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4049929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:00:48.4050080Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:00:48.4050085Z 2025-09-07T08:00:48.4050218Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4050757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4050872Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4051300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4051383Z layer_outputs = layer( 2025-09-07T08:00:48.4051805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4051897Z attn_outputs = self.attention( 2025-09-07T08:00:48.4052319Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4052436Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4052858Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T08:00:48.4053160Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:00:48.4053709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:00:48.4053946Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:00:48.4053977Z 2025-09-07T08:00:48.4054116Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4054652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4054767Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4055191Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4055308Z layer_outputs = layer( 2025-09-07T08:00:48.4055731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4055820Z attn_outputs = self.attention( 2025-09-07T08:00:48.4056248Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T08:00:48.4056442Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T08:00:48.4056865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T08:00:48.4056969Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.4056974Z 2025-09-07T08:00:48.4057106Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4057642Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4057758Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4058185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4058268Z layer_outputs = layer( 2025-09-07T08:00:48.4058690Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.4058859Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.4059282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.4059377Z return apply_chunking_to_forward( 2025-09-07T08:00:48.4059780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.4059873Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.4060328Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:00:48.4060433Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.4060861Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T08:00:48.4060965Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.4060970Z 2025-09-07T08:00:48.4061106Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4061640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4061751Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4062178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4062258Z layer_outputs = layer( 2025-09-07T08:00:48.4062681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.4062922Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.4063353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.4063476Z return apply_chunking_to_forward( 2025-09-07T08:00:48.4063878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.4063971Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.4064419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:00:48.4064551Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.4064972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T08:00:48.4065075Z hidden_states = self.act_fn(hidden_states) 2025-09-07T08:00:48.4065084Z 2025-09-07T08:00:48.4065220Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4065826Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4065945Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4066365Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4066446Z layer_outputs = layer( 2025-09-07T08:00:48.4066865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.4067030Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.4067452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.4067547Z return apply_chunking_to_forward( 2025-09-07T08:00:48.4067951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.4068045Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.4068492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T08:00:48.4068587Z return self.output(hidden_states) 2025-09-07T08:00:48.4069007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T08:00:48.4069111Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.4069117Z 2025-09-07T08:00:48.4069252Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4069792Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4069906Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4070334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4070414Z layer_outputs = layer( 2025-09-07T08:00:48.4070837Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4070936Z attn_outputs = self.attention( 2025-09-07T08:00:48.4071360Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4071482Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4071902Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 540, in forward 2025-09-07T08:00:48.4072136Z query_key_vectors = self.query_key(hidden_states) 2025-09-07T08:00:48.4072203Z 2025-09-07T08:00:48.4072349Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4072939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4073057Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4073478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4073557Z layer_outputs = layer( 2025-09-07T08:00:48.4073979Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4074120Z attn_outputs = self.attention( 2025-09-07T08:00:48.4074548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4074668Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4075089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:00:48.4075293Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:00:48.4075738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T08:00:48.4075953Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T08:00:48.4075958Z 2025-09-07T08:00:48.4076091Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4076632Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4076747Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4077175Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4077256Z layer_outputs = layer( 2025-09-07T08:00:48.4077676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4077767Z attn_outputs = self.attention( 2025-09-07T08:00:48.4078190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4078307Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4078731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:00:48.4078929Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:00:48.4079379Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T08:00:48.4079577Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T08:00:48.4079584Z 2025-09-07T08:00:48.4079722Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4080255Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4080372Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4080796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4080878Z layer_outputs = layer( 2025-09-07T08:00:48.4081302Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4081461Z attn_outputs = self.attention( 2025-09-07T08:00:48.4081930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4082074Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4082496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:00:48.4082698Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:00:48.4083140Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T08:00:48.4083287Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T08:00:48.4083292Z 2025-09-07T08:00:48.4083427Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4083965Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4084078Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4084506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4084585Z layer_outputs = layer( 2025-09-07T08:00:48.4085007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4085099Z attn_outputs = self.attention( 2025-09-07T08:00:48.4085519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4085639Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4086058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:00:48.4086335Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:00:48.4086903Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:00:48.4087039Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:00:48.4087493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T08:00:48.4087704Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T08:00:48.4087712Z 2025-09-07T08:00:48.4087813Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4087947Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4088485Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4088605Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4089029Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4089111Z layer_outputs = layer( 2025-09-07T08:00:48.4089534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4089623Z attn_outputs = self.attention( 2025-09-07T08:00:48.4090047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4090165Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4090587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:00:48.4090927Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:00:48.4091502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:00:48.4091661Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:00:48.4092117Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T08:00:48.4092229Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T08:00:48.4092234Z 2025-09-07T08:00:48.4092331Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4092453Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4092586Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4093128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4093243Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4093674Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4093759Z layer_outputs = layer( 2025-09-07T08:00:48.4094183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4094278Z attn_outputs = self.attention( 2025-09-07T08:00:48.4094698Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4094817Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4095240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.4095374Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.4095800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:00:48.4096102Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.4096553Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:00:48.4096757Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:00:48.4096762Z 2025-09-07T08:00:48.4096859Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4096962Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4097096Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4097638Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4097755Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4098186Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4098266Z layer_outputs = layer( 2025-09-07T08:00:48.4098689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4098779Z attn_outputs = self.attention( 2025-09-07T08:00:48.4099203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4099322Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4099744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.4099917Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.4100373Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:00:48.4100696Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.4101150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:00:48.4101240Z return torch.cat(slices, dim=3) 2025-09-07T08:00:48.4101245Z 2025-09-07T08:00:48.4101345Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4101479Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4102039Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4102160Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4102584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4102670Z layer_outputs = layer( 2025-09-07T08:00:48.4103092Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4103179Z attn_outputs = self.attention( 2025-09-07T08:00:48.4103605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4103721Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4104145Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 541, in forward 2025-09-07T08:00:48.4104247Z value_vectors = self.value(hidden_states) 2025-09-07T08:00:48.4104251Z 2025-09-07T08:00:48.4104390Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4104927Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4105042Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4105468Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4105547Z layer_outputs = layer( 2025-09-07T08:00:48.4105974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4106061Z attn_outputs = self.attention( 2025-09-07T08:00:48.4106488Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4106604Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4107025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.4107164Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.4107585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:00:48.4107846Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.4108294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:00:48.4108497Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:00:48.4108508Z 2025-09-07T08:00:48.4108607Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4108741Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4109359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4109504Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4109933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4110012Z layer_outputs = layer( 2025-09-07T08:00:48.4110435Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4110527Z attn_outputs = self.attention( 2025-09-07T08:00:48.4110951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4111119Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4111542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.4111679Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.4112099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:00:48.4112356Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.4112810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:00:48.4112898Z return torch.cat(slices, dim=3) 2025-09-07T08:00:48.4112903Z 2025-09-07T08:00:48.4113040Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4113576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4113695Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4114123Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4114203Z layer_outputs = layer( 2025-09-07T08:00:48.4114627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4114715Z attn_outputs = self.attention( 2025-09-07T08:00:48.4115137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4115252Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4115673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T08:00:48.4115968Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T08:00:48.4116448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T08:00:48.4116570Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T08:00:48.4116575Z 2025-09-07T08:00:48.4116707Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4117245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4117359Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4117783Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4117869Z layer_outputs = layer( 2025-09-07T08:00:48.4118292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4118446Z attn_outputs = self.attention( 2025-09-07T08:00:48.4118934Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4119086Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4119508Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.4119642Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.4120064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T08:00:48.4120301Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:00:48.4120306Z 2025-09-07T08:00:48.4120444Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4120990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4121104Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4121535Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4121614Z layer_outputs = layer( 2025-09-07T08:00:48.4122041Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4122130Z attn_outputs = self.attention( 2025-09-07T08:00:48.4122555Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4122673Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4123093Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.4123231Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.4123650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T08:00:48.4123934Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:00:48.4124385Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:00:48.4124590Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:00:48.4124595Z 2025-09-07T08:00:48.4124695Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4124828Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4125374Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4125489Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4125920Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4126002Z layer_outputs = layer( 2025-09-07T08:00:48.4126425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4126517Z attn_outputs = self.attention( 2025-09-07T08:00:48.4126941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4127061Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4127480Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.4127674Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.4128135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:00:48.4128313Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:00:48.4128318Z 2025-09-07T08:00:48.4128456Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4128992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4129106Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4129530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4129634Z layer_outputs = layer( 2025-09-07T08:00:48.4130059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4130148Z attn_outputs = self.attention( 2025-09-07T08:00:48.4130571Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4130691Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4131111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:00:48.4131241Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:00:48.4131659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:00:48.4131814Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:00:48.4131819Z 2025-09-07T08:00:48.4131915Z cudagraph partition due to non gpu ops 2025-09-07T08:00:48.4132052Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4132590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4132709Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4133132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4133209Z layer_outputs = layer( 2025-09-07T08:00:48.4133633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4133721Z attn_outputs = self.attention( 2025-09-07T08:00:48.4134146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:00:48.4134262Z self_attention_outputs = self.self_attention( 2025-09-07T08:00:48.4134681Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T08:00:48.4134983Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:00:48.4135471Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:00:48.4135670Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:00:48.4135675Z 2025-09-07T08:00:48.4135811Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4136352Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4136470Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4136955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4137076Z layer_outputs = layer( 2025-09-07T08:00:48.4137503Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:00:48.4137627Z attn_outputs = self.attention( 2025-09-07T08:00:48.4138051Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T08:00:48.4138244Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T08:00:48.4138666Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T08:00:48.4138805Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.4138809Z 2025-09-07T08:00:48.4138947Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4139487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4139604Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4140026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4140105Z layer_outputs = layer( 2025-09-07T08:00:48.4140525Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.4140687Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.4141111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.4141206Z return apply_chunking_to_forward( 2025-09-07T08:00:48.4141611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.4141707Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.4142159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:00:48.4142265Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.4142687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T08:00:48.4142790Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.4142795Z 2025-09-07T08:00:48.4142929Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4143470Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4143584Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4144011Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4144096Z layer_outputs = layer( 2025-09-07T08:00:48.4144520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.4144685Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.4145106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.4145198Z return apply_chunking_to_forward( 2025-09-07T08:00:48.4145603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.4145695Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.4146203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:00:48.4146348Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.4146774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T08:00:48.4146926Z hidden_states = self.act_fn(hidden_states) 2025-09-07T08:00:48.4146931Z 2025-09-07T08:00:48.4147065Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4147604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4147715Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4148189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4148271Z layer_outputs = layer( 2025-09-07T08:00:48.4148698Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.4148860Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.4149283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:00:48.4149376Z return apply_chunking_to_forward( 2025-09-07T08:00:48.4149778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:48.4149873Z return forward_fn(*input_tensors) 2025-09-07T08:00:48.4150321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T08:00:48.4150415Z return self.output(hidden_states) 2025-09-07T08:00:48.4150843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T08:00:48.4150948Z hidden_states = self.dense(hidden_states) 2025-09-07T08:00:48.4150952Z 2025-09-07T08:00:48.4151091Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4151627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4151742Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4152166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:00:48.4152243Z layer_outputs = layer( 2025-09-07T08:00:48.4152670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:00:48.4152833Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:00:48.4152837Z 2025-09-07T08:00:48.4152978Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:48.4153513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:00:48.4153630Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:00:48.4154053Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1753, in forward 2025-09-07T08:00:48.4154191Z return torch.cat([attn_output, hidden_states], dim=-1) 2025-09-07T08:00:48.4154197Z 2025-09-07T08:00:52.7608573Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:00:52.7609421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 2495, in torch_dynamo_resume_in_forward_at_2481 2025-09-07T08:00:52.7610168Z logits = self.lm_head(sequence_output) 2025-09-07T08:00:52.7611080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1886, in forward 2025-09-07T08:00:52.7611948Z return apply_chunking_to_forward(self.forward_chunk, self.chunk_size_lm_head, self.seq_len_dim, hidden_states) 2025-09-07T08:00:52.7612887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:00:52.7613486Z return forward_fn(*input_tensors) 2025-09-07T08:00:52.7614120Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/transformers/models/reformer/modeling_reformer.py", line 1889, in forward_chunk 2025-09-07T08:00:52.7614778Z hidden_states = self.decoder(hidden_states) 2025-09-07T08:00:52.7615094Z 2025-09-07T08:00:52.9462597Z 2025-09-07T08:00:53.1322341Z running benchmark: 0% 0/30 [00:00> $GITHUB_ENV 2025-09-07T08:19:57.9891687Z echo "DEVICE_TYPE=$DEVICE_TYPE" >> $GITHUB_ENV 2025-09-07T08:19:57.9901366Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:19:57.9901693Z env: 2025-09-07T08:19:57.9901874Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:19:57.9902271Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:19:57.9902688Z ##[endgroup] 2025-09-07T08:19:57.9930908Z + [[ -n '' ]] 2025-09-07T08:19:57.9931221Z + python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-09-07T08:19:58.2089675Z Defaulting to user installation because normal site-packages is not writeable 2025-09-07T08:19:59.1212983Z Collecting boto3==1.35.33 2025-09-07T08:19:59.1359532Z Downloading boto3-1.35.33-py3-none-any.whl (139 kB) 2025-09-07T08:19:59.3885674Z Collecting psutil==7.0.0 2025-09-07T08:19:59.3916629Z Downloading psutil-7.0.0-cp36-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (279 kB) 2025-09-07T08:19:59.4179400Z Collecting pynvml==12.0.0 2025-09-07T08:19:59.4208972Z Downloading pynvml-12.0.0-py3-none-any.whl (26 kB) 2025-09-07T08:19:59.4602822Z Collecting s3transfer<0.11.0,>=0.10.0 2025-09-07T08:19:59.4634665Z Downloading s3transfer-0.10.4-py3-none-any.whl (83 kB) 2025-09-07T08:20:00.4786572Z Collecting botocore<1.36.0,>=1.35.33 2025-09-07T08:20:00.4817726Z Downloading botocore-1.35.99-py3-none-any.whl (13.3 MB) 2025-09-07T08:20:00.5901911Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.33) (0.10.0) 2025-09-07T08:20:00.6287568Z Collecting nvidia-ml-py<13.0.0a0,>=12.0.0 2025-09-07T08:20:00.6320632Z Downloading nvidia_ml_py-12.575.51-py3-none-any.whl (47 kB) 2025-09-07T08:20:00.6419484Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.25.10) 2025-09-07T08:20:00.6423500Z Requirement already satisfied: python-dateutil<3.0.0,>=2.1 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (2.8.1) 2025-09-07T08:20:00.8800516Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil<3.0.0,>=2.1->botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.15.0) 2025-09-07T08:20:00.9789608Z Installing collected packages: botocore, s3transfer, nvidia-ml-py, pynvml, psutil, boto3 2025-09-07T08:20:01.6814281Z Successfully installed boto3-1.35.33 botocore-1.35.99 nvidia-ml-py-12.575.51 psutil-7.0.0 pynvml-12.0.0 s3transfer-0.10.4 2025-09-07T08:20:01.7632685Z + DEVICE_NAME= 2025-09-07T08:20:01.7632879Z + DEVICE_TYPE= 2025-09-07T08:20:01.7633086Z + command -v nvidia-smi 2025-09-07T08:20:01.7633303Z + command -v rocminfo 2025-09-07T08:20:01.7633507Z + echo DEVICE_NAME= 2025-09-07T08:20:01.7634792Z + echo DEVICE_TYPE= 2025-09-07T08:20:01.7656176Z ##[group]Run set -eux 2025-09-07T08:20:01.7656427Z set -eux 2025-09-07T08:20:01.7656613Z  2025-09-07T08:20:01.7656824Z if [[ -z "${GITHUB_TOKEN}" ]]; then 2025-09-07T08:20:01.7657131Z  echo "Missing github-token input" 2025-09-07T08:20:01.7657403Z  exit 1 2025-09-07T08:20:01.7657588Z fi 2025-09-07T08:20:01.7665529Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:01.7665868Z env: 2025-09-07T08:20:01.7666046Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:01.7666441Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:01.7666863Z DEVICE_NAME: 2025-09-07T08:20:01.7667067Z DEVICE_TYPE: 2025-09-07T08:20:01.7667463Z GITHUB_TOKEN: *** 2025-09-07T08:20:01.7667654Z ##[endgroup] 2025-09-07T08:20:01.7691584Z + [[ -z *** ]] 2025-09-07T08:20:01.7730345Z ##[group]Run pytorch/test-infra/.github/actions/get-workflow-job-id@main 2025-09-07T08:20:01.7730707Z with: 2025-09-07T08:20:01.7731030Z github-token: *** 2025-09-07T08:20:01.7731219Z env: 2025-09-07T08:20:01.7731387Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:01.7731785Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:01.7732206Z DEVICE_NAME: 2025-09-07T08:20:01.7732387Z DEVICE_TYPE: 2025-09-07T08:20:01.7732562Z ##[endgroup] 2025-09-07T08:20:01.7745606Z ##[group]Run set -eux 2025-09-07T08:20:01.7745823Z set -eux 2025-09-07T08:20:01.7746007Z  2025-09-07T08:20:01.7746406Z python3 "${GITHUB_ACTION_PATH}/../../scripts/get_workflow_job_id.py" "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-09-07T08:20:01.7752438Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:01.7752766Z env: 2025-09-07T08:20:01.7752941Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:01.7753341Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:01.7753759Z DEVICE_NAME: 2025-09-07T08:20:01.7753941Z DEVICE_TYPE: 2025-09-07T08:20:01.7754292Z GITHUB_TOKEN: *** 2025-09-07T08:20:01.7754487Z ##[endgroup] 2025-09-07T08:20:01.7776681Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/get-workflow-job-id/../../scripts/get_workflow_job_id.py 17525321311 i-0c7ee175fda0aa6a4 2025-09-07T08:20:03.0602213Z setting job-id=49775354812 2025-09-07T08:20:03.0602784Z setting job-name=linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal) 2025-09-07T08:20:03.0698398Z ##[group]Run set -eux 2025-09-07T08:20:03.0698624Z set -eux 2025-09-07T08:20:03.0698810Z  2025-09-07T08:20:03.0699281Z if [[ -n "" ]]; then 2025-09-07T08:20:03.0699520Z  source "" 2025-09-07T08:20:03.0699714Z fi 2025-09-07T08:20:03.0699882Z  2025-09-07T08:20:03.0700210Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_metadata.py" \ 2025-09-07T08:20:03.0700657Z  --schema-version "${SCHEMA_VERSION}" \ 2025-09-07T08:20:03.0700949Z  --repo "${REPO}" \ 2025-09-07T08:20:03.0701303Z  --head-branch "${HEAD_BRANCH}" \ 2025-09-07T08:20:03.0701586Z  --head-sha "${HEAD_SHA}" \ 2025-09-07T08:20:03.0701866Z  --workflow-id "${WORKFLOW_RUN_ID}" \ 2025-09-07T08:20:03.0702169Z  --run-attempt "${RUN_ATTEMPT}" \ 2025-09-07T08:20:03.0702444Z  --job-id "${JOB_ID}" \ 2025-09-07T08:20:03.0702696Z  --job-name "${JOB_NAME}" 2025-09-07T08:20:03.0708788Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:03.0709110Z env: 2025-09-07T08:20:03.0709294Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:03.0709702Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:03.0710123Z DEVICE_NAME: 2025-09-07T08:20:03.0710304Z DEVICE_TYPE: 2025-09-07T08:20:03.0710488Z SCHEMA_VERSION: v3 2025-09-07T08:20:03.0710692Z REPO: pytorch/pytorch 2025-09-07T08:20:03.0710904Z HEAD_BRANCH: refs/heads/main 2025-09-07T08:20:03.0711175Z HEAD_SHA: 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T08:20:03.0711470Z WORKFLOW_RUN_ID: 17525321311 2025-09-07T08:20:03.0711691Z RUN_ATTEMPT: 1 2025-09-07T08:20:03.0711875Z JOB_ID: 49775354812 2025-09-07T08:20:03.0712373Z JOB_NAME: linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal) 2025-09-07T08:20:03.0712910Z ##[endgroup] 2025-09-07T08:20:03.0735253Z + [[ -n '' ]] 2025-09-07T08:20:03.0737243Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_metadata.py --schema-version v3 --repo pytorch/pytorch --head-branch refs/heads/main --head-sha 93fb23d6fae7c4e82c4239a1033e522088742634 --workflow-id 17525321311 --run-attempt 1 --job-id 49775354812 --job-name 'linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal)' 2025-09-07T08:20:03.1029399Z ##[group]Run set -eux 2025-09-07T08:20:03.1029625Z set -eux 2025-09-07T08:20:03.1029812Z  2025-09-07T08:20:03.1029988Z if [[ -n "" ]]; then 2025-09-07T08:20:03.1030209Z  source "" 2025-09-07T08:20:03.1030401Z fi 2025-09-07T08:20:03.1030568Z  2025-09-07T08:20:03.1030901Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_runners_info.py" 2025-09-07T08:20:03.1037094Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:03.1037416Z env: 2025-09-07T08:20:03.1037588Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:03.1037991Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:03.1038419Z DEVICE_NAME: 2025-09-07T08:20:03.1038601Z DEVICE_TYPE: 2025-09-07T08:20:03.1038777Z ##[endgroup] 2025-09-07T08:20:03.1059569Z + [[ -n '' ]] 2025-09-07T08:20:03.1060256Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_runners_info.py 2025-09-07T08:20:03.1461295Z INFO:root:Fail to import torch to get the device name 2025-09-07T08:20:03.1557954Z ##[group]Run set -eux 2025-09-07T08:20:03.1558175Z set -eux 2025-09-07T08:20:03.1558357Z  2025-09-07T08:20:03.1558556Z # TODO (huydhn): Implement this part 2025-09-07T08:20:03.1558878Z echo "dependencies={}" >> "${GITHUB_OUTPUT}" 2025-09-07T08:20:03.1564732Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:03.1565047Z env: 2025-09-07T08:20:03.1565217Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:03.1565981Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:03.1566409Z DEVICE_NAME: 2025-09-07T08:20:03.1566590Z DEVICE_TYPE: 2025-09-07T08:20:03.1566771Z ##[endgroup] 2025-09-07T08:20:03.1588265Z + echo 'dependencies={}' 2025-09-07T08:20:03.1606568Z ##[group]Run set -eux 2025-09-07T08:20:03.1606797Z set -eux 2025-09-07T08:20:03.1606981Z  2025-09-07T08:20:03.1607300Z if [[ -n "" ]]; then 2025-09-07T08:20:03.1607520Z  source "" 2025-09-07T08:20:03.1607715Z fi 2025-09-07T08:20:03.1607884Z  2025-09-07T08:20:03.1608103Z if [[ ! -d "${BENCHMARK_RESULTS_DIR}" ]]; then 2025-09-07T08:20:03.1608473Z  echo "${BENCHMARK_RESULTS_DIR} does not exist, skipping" 2025-09-07T08:20:03.1608891Z  # We don't want the job to fail if the directory doesn't exist 2025-09-07T08:20:03.1609220Z  exit 0 2025-09-07T08:20:03.1609398Z fi 2025-09-07T08:20:03.1609572Z  2025-09-07T08:20:03.1609765Z if [[ "${DRY_RUN}" == "true" ]]; then 2025-09-07T08:20:03.1610177Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-09-07T08:20:03.1610660Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-09-07T08:20:03.1611031Z  --metadata "${BENCHMARK_METADATA}" \ 2025-09-07T08:20:03.1611332Z  --runners "${RUNNER_INFO}" \ 2025-09-07T08:20:03.1611630Z  --dependencies "${DEPENDENCIES}" \ 2025-09-07T08:20:03.1611909Z  --dry-run 2025-09-07T08:20:03.1612112Z else 2025-09-07T08:20:03.1612432Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-09-07T08:20:03.1612908Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-09-07T08:20:03.1613275Z  --metadata "${BENCHMARK_METADATA}" \ 2025-09-07T08:20:03.1613574Z  --runners "${RUNNER_INFO}" \ 2025-09-07T08:20:03.1613872Z  --dependencies "${DEPENDENCIES}" 2025-09-07T08:20:03.1614142Z fi 2025-09-07T08:20:03.1619871Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:03.1620187Z env: 2025-09-07T08:20:03.1620357Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:03.1620754Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:03.1621176Z DEVICE_NAME: 2025-09-07T08:20:03.1621363Z DEVICE_TYPE: 2025-09-07T08:20:03.1621573Z BENCHMARK_RESULTS_DIR: test/test-reports 2025-09-07T08:20:03.1621829Z DRY_RUN: false 2025-09-07T08:20:03.1623089Z BENCHMARK_METADATA: {"timestamp": 1757233203, "schema_version": "v3", "name": "linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "93fb23d6fae7c4e82c4239a1033e522088742634", "workflow_id": 17525321311, "run_attempt": 1, "job_id": 49775354812} 2025-09-07T08:20:03.1624722Z RUNNER_INFO: [{"cpu_info": "aarch64", "cpu_count": 64, "avail_mem_in_gb": 251, "extra_info": {"hostname": "ip-10-0-72-124.ec2.internal"}, "name": "", "type": ""}] 2025-09-07T08:20:03.1625287Z DEPENDENCIES: {} 2025-09-07T08:20:03.1625474Z ##[endgroup] 2025-09-07T08:20:03.1646179Z + [[ -n '' ]] 2025-09-07T08:20:03.1646390Z + [[ ! -d test/test-reports ]] 2025-09-07T08:20:03.1646624Z + [[ false == \t\r\u\e ]] 2025-09-07T08:20:03.1649310Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/upload_benchmark_results.py --benchmark-results-dir test/test-reports --metadata '{"timestamp": 1757233203, "schema_version": "v3", "name": "linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "93fb23d6fae7c4e82c4239a1033e522088742634", "workflow_id": 17525321311, "run_attempt": 1, "job_id": 49775354812}' --runners '[{"cpu_info": "aarch64", "cpu_count": 64, "avail_mem_in_gb": 251, "extra_info": {"hostname": "ip-10-0-72-124.ec2.internal"}, "name": "", "type": ""}]' --dependencies '{}' 2025-09-07T08:20:03.2908013Z INFO:root:Upload test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_aarch64_accuracy.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525321311/49775354812/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_aarch64_accuracy.json 2025-09-07T08:20:03.3158363Z INFO:botocore.credentials:Found credentials from IAM Role: gh-ci-github-action-runners-runner-role 2025-09-07T08:20:03.4686804Z INFO:root:Upload test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_aarch64_accuracy.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525321311/49775354812/inductor_dynamic_torchbench_bfloat16_inference_cpu_aarch64_accuracy.json 2025-09-07T08:20:03.5492827Z INFO:root:Upload test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_aarch64_accuracy.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525321311/49775354812/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_aarch64_accuracy.json 2025-09-07T08:20:03.6303543Z INFO:root:Upload test/test-reports/inductor_export_torchbench_bfloat16_inference_cpu_aarch64_accuracy.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525321311/49775354812/inductor_export_torchbench_bfloat16_inference_cpu_aarch64_accuracy.json 2025-09-07T08:20:03.7362861Z INFO:root:Upload test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_aarch64_accuracy.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525321311/49775354812/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_aarch64_accuracy.json 2025-09-07T08:20:03.8202543Z INFO:root:Upload test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_aarch64_performance.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525321311/49775354812/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_aarch64_performance.json 2025-09-07T08:20:03.9028263Z INFO:root:Upload test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525321311/49775354812/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.json 2025-09-07T08:20:03.9825296Z INFO:root:Upload test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_aarch64_performance.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525321311/49775354812/inductor_dynamic_torchbench_bfloat16_inference_cpu_aarch64_performance.json 2025-09-07T08:20:04.0604362Z INFO:root:Upload test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525321311/49775354812/inductor_dynamic_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.json 2025-09-07T08:20:04.1459350Z INFO:root:Upload test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_aarch64_performance.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525321311/49775354812/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_aarch64_performance.json 2025-09-07T08:20:04.2219030Z INFO:root:Upload test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525321311/49775354812/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.json 2025-09-07T08:20:04.2957592Z INFO:root:Upload test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_aarch64_performance.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525321311/49775354812/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_aarch64_performance.json 2025-09-07T08:20:04.3761388Z INFO:root:Upload test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525321311/49775354812/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.json 2025-09-07T08:20:04.4949190Z ##[group]Run cat test/**/*_toprint.log || true 2025-09-07T08:20:04.4949574Z cat test/**/*_toprint.log || true 2025-09-07T08:20:04.4956088Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:04.4956406Z env: 2025-09-07T08:20:04.4956583Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:04.4957109Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:04.4957532Z DEVICE_NAME: 2025-09-07T08:20:04.4957718Z DEVICE_TYPE: 2025-09-07T08:20:04.4957901Z ##[endgroup] 2025-09-07T08:20:04.5033071Z cat: 'test/**/*_toprint.log': No such file or directory 2025-09-07T08:20:04.5147116Z Prepare all required actions 2025-09-07T08:20:04.5147520Z Getting action download info 2025-09-07T08:20:04.6784421Z Download action repository 'seemethere/upload-artifact-s3@v5' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-09-07T08:20:04.8514952Z Download action repository 'actions/upload-artifact@v4' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-09-07T08:20:05.2077703Z ##[group]Run ./.github/actions/upload-test-artifacts 2025-09-07T08:20:05.2078002Z with: 2025-09-07T08:20:05.2078379Z file-suffix: test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812 2025-09-07T08:20:05.2078840Z s3-bucket: gha-artifacts 2025-09-07T08:20:05.2079065Z env: 2025-09-07T08:20:05.2079238Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:05.2079628Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:05.2080050Z DEVICE_NAME: 2025-09-07T08:20:05.2080232Z DEVICE_TYPE: 2025-09-07T08:20:05.2080406Z ##[endgroup] 2025-09-07T08:20:05.2100890Z ##[group]Run # Remove any previous test jsons if they exist 2025-09-07T08:20:05.2101286Z # Remove any previous test jsons if they exist 2025-09-07T08:20:05.2101597Z rm -f test-jsons-*.zip 2025-09-07T08:20:05.2101970Z zip -r "test-jsons-${FILE_SUFFIX}.zip" test/test-reports -i '*.json' 2025-09-07T08:20:05.2108482Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:05.2108805Z env: 2025-09-07T08:20:05.2108979Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:05.2109373Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:05.2109804Z DEVICE_NAME: 2025-09-07T08:20:05.2109989Z DEVICE_TYPE: 2025-09-07T08:20:05.2110373Z FILE_SUFFIX: test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812 2025-09-07T08:20:05.2110810Z ##[endgroup] 2025-09-07T08:20:05.2244290Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_aarch64_accuracy.json (deflated 98%) 2025-09-07T08:20:05.2252665Z adding: test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_aarch64_accuracy.json (deflated 98%) 2025-09-07T08:20:05.2261125Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_aarch64_accuracy.json (deflated 98%) 2025-09-07T08:20:05.2269601Z adding: test/test-reports/inductor_export_torchbench_bfloat16_inference_cpu_aarch64_accuracy.json (deflated 98%) 2025-09-07T08:20:05.2276463Z adding: test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_aarch64_accuracy.json (deflated 98%) 2025-09-07T08:20:05.2288932Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_aarch64_performance.json (deflated 98%) 2025-09-07T08:20:05.2316278Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.json (deflated 98%) 2025-09-07T08:20:05.2328616Z adding: test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_aarch64_performance.json (deflated 98%) 2025-09-07T08:20:05.2356113Z adding: test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.json (deflated 98%) 2025-09-07T08:20:05.2368760Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_aarch64_performance.json (deflated 98%) 2025-09-07T08:20:05.2395817Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.json (deflated 98%) 2025-09-07T08:20:05.2405843Z adding: test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_aarch64_performance.json (deflated 98%) 2025-09-07T08:20:05.2420196Z adding: test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.json (deflated 98%) 2025-09-07T08:20:05.2441053Z ##[group]Run # Remove any previous test reports if they exist 2025-09-07T08:20:05.2441684Z # Remove any previous test reports if they exist 2025-09-07T08:20:05.2442016Z rm -f test-reports-*.zip 2025-09-07T08:20:05.2442425Z zip -r "test-reports-${FILE_SUFFIX}.zip" test/test-reports -i '*.xml' -i '*.csv' 2025-09-07T08:20:05.2448179Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:05.2448504Z env: 2025-09-07T08:20:05.2448678Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:05.2449077Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:05.2449498Z DEVICE_NAME: 2025-09-07T08:20:05.2449682Z DEVICE_TYPE: 2025-09-07T08:20:05.2450067Z FILE_SUFFIX: test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812 2025-09-07T08:20:05.2450508Z ##[endgroup] 2025-09-07T08:20:05.2497800Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_aarch64_accuracy.csv (deflated 47%) 2025-09-07T08:20:05.2498656Z adding: test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_aarch64_accuracy.csv (deflated 47%) 2025-09-07T08:20:05.2499472Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_aarch64_accuracy.csv (deflated 47%) 2025-09-07T08:20:05.2500284Z adding: test/test-reports/inductor_export_torchbench_bfloat16_inference_cpu_aarch64_accuracy.csv (deflated 48%) 2025-09-07T08:20:05.2501103Z adding: test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_aarch64_accuracy.csv (deflated 54%) 2025-09-07T08:20:05.2501949Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_aarch64_performance.csv (deflated 46%) 2025-09-07T08:20:05.2502883Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.csv (deflated 48%) 2025-09-07T08:20:05.2503795Z adding: test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_aarch64_performance.csv (deflated 45%) 2025-09-07T08:20:05.2504695Z adding: test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.csv (deflated 47%) 2025-09-07T08:20:05.2505602Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_aarch64_performance.csv (deflated 45%) 2025-09-07T08:20:05.2506717Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.csv (deflated 48%) 2025-09-07T08:20:05.2507639Z adding: test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_aarch64_performance.csv (deflated 45%) 2025-09-07T08:20:05.2508561Z adding: test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_aarch64_performance_compilation_metrics.csv (deflated 46%) 2025-09-07T08:20:05.2525846Z ##[group]Run # Remove any previous usage logs if they exist 2025-09-07T08:20:05.2526238Z # Remove any previous usage logs if they exist 2025-09-07T08:20:05.2526547Z rm -f logs-*.zip 2025-09-07T08:20:05.2526847Z zip "logs-${FILE_SUFFIX}.zip" 'usage_log.txt' || true 2025-09-07T08:20:05.2527282Z zip -r "logs-${FILE_SUFFIX}.zip" test/test-reports -i '*.log' || true 2025-09-07T08:20:05.2532875Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:05.2533319Z env: 2025-09-07T08:20:05.2533497Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:05.2533891Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:05.2534315Z DEVICE_NAME: 2025-09-07T08:20:05.2534500Z DEVICE_TYPE: 2025-09-07T08:20:05.2534884Z FILE_SUFFIX: test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812 2025-09-07T08:20:05.2535420Z ##[endgroup] 2025-09-07T08:20:05.2573484Z zip warning: name not matched: usage_log.txt 2025-09-07T08:20:05.2573699Z 2025-09-07T08:20:05.2574052Z zip error: Nothing to do! (logs-test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812.zip) 2025-09-07T08:20:05.2583868Z zip warning: zip file empty 2025-09-07T08:20:05.2603768Z ##[group]Run # Remove any previous debugging artifacts if they exist 2025-09-07T08:20:05.2624094Z # Remove any previous debugging artifacts if they exist 2025-09-07T08:20:05.2624483Z rm -f debug-*.zip 2025-09-07T08:20:05.2624732Z if [ -d 'test/debug' ]; then 2025-09-07T08:20:05.2625043Z  zip -r "debug-${FILE_SUFFIX}.zip" test/debug 2025-09-07T08:20:05.2625328Z fi 2025-09-07T08:20:05.2630960Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:05.2631286Z env: 2025-09-07T08:20:05.2631467Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:05.2631866Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:05.2632302Z DEVICE_NAME: 2025-09-07T08:20:05.2632489Z DEVICE_TYPE: 2025-09-07T08:20:05.2632879Z FILE_SUFFIX: test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812 2025-09-07T08:20:05.2633321Z ##[endgroup] 2025-09-07T08:20:05.2722414Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-09-07T08:20:05.2722689Z with: 2025-09-07T08:20:05.2722861Z s3-bucket: gha-artifacts 2025-09-07T08:20:05.2723130Z s3-prefix: pytorch/pytorch/17525321311/1/artifact 2025-09-07T08:20:05.2723423Z retention-days: 14 2025-09-07T08:20:05.2723631Z if-no-files-found: warn 2025-09-07T08:20:05.2723849Z path: test-jsons-*.zip 2025-09-07T08:20:05.2724056Z name: artifact 2025-09-07T08:20:05.2724243Z region: us-east-1 2025-09-07T08:20:05.2724417Z env: 2025-09-07T08:20:05.2724583Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:05.2724978Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:05.2725403Z DEVICE_NAME: 2025-09-07T08:20:05.2725583Z DEVICE_TYPE: 2025-09-07T08:20:05.2725754Z ##[endgroup] 2025-09-07T08:20:05.5556505Z NOTE: s3-prefix specified, ignoring name parameter 2025-09-07T08:20:05.5556902Z With the provided path, there will be 1 file uploaded 2025-09-07T08:20:05.5557317Z Uploading to s3 prefix: pytorch/pytorch/17525321311/1/artifact 2025-09-07T08:20:05.5596781Z Starting upload of test-jsons-test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812.zip 2025-09-07T08:20:05.6882648Z Finished upload of test-jsons-test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812.zip 2025-09-07T08:20:05.7086766Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-09-07T08:20:05.7087046Z with: 2025-09-07T08:20:05.7087227Z s3-bucket: gha-artifacts 2025-09-07T08:20:05.7087498Z s3-prefix: pytorch/pytorch/17525321311/1/artifact 2025-09-07T08:20:05.7087791Z retention-days: 14 2025-09-07T08:20:05.7088012Z if-no-files-found: error 2025-09-07T08:20:05.7088243Z path: test-reports-*.zip 2025-09-07T08:20:05.7088459Z name: artifact 2025-09-07T08:20:05.7088647Z region: us-east-1 2025-09-07T08:20:05.7088825Z env: 2025-09-07T08:20:05.7088999Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:05.7089406Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:05.7089834Z DEVICE_NAME: 2025-09-07T08:20:05.7090015Z DEVICE_TYPE: 2025-09-07T08:20:05.7090193Z ##[endgroup] 2025-09-07T08:20:05.9681287Z NOTE: s3-prefix specified, ignoring name parameter 2025-09-07T08:20:05.9681907Z With the provided path, there will be 1 file uploaded 2025-09-07T08:20:05.9682300Z Uploading to s3 prefix: pytorch/pytorch/17525321311/1/artifact 2025-09-07T08:20:05.9721773Z Starting upload of test-reports-test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812.zip 2025-09-07T08:20:06.1157476Z Finished upload of test-reports-test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812.zip 2025-09-07T08:20:06.1363840Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-09-07T08:20:06.1364118Z with: 2025-09-07T08:20:06.1364297Z s3-bucket: gha-artifacts 2025-09-07T08:20:06.1364559Z s3-prefix: pytorch/pytorch/17525321311/1/artifact 2025-09-07T08:20:06.1364850Z retention-days: 14 2025-09-07T08:20:06.1365361Z if-no-files-found: ignore 2025-09-07T08:20:06.1365596Z path: logs-*.zip 2025-09-07T08:20:06.1365780Z name: artifact 2025-09-07T08:20:06.1365965Z region: us-east-1 2025-09-07T08:20:06.1366152Z env: 2025-09-07T08:20:06.1366323Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:06.1366716Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:06.1367135Z DEVICE_NAME: 2025-09-07T08:20:06.1367316Z DEVICE_TYPE: 2025-09-07T08:20:06.1367490Z ##[endgroup] 2025-09-07T08:20:06.3956534Z NOTE: s3-prefix specified, ignoring name parameter 2025-09-07T08:20:06.3956910Z With the provided path, there will be 1 file uploaded 2025-09-07T08:20:06.3957311Z Uploading to s3 prefix: pytorch/pytorch/17525321311/1/artifact 2025-09-07T08:20:06.3997241Z Starting upload of logs-test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812.zip 2025-09-07T08:20:06.5262222Z Finished upload of logs-test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812.zip 2025-09-07T08:20:06.5466084Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-09-07T08:20:06.5466364Z with: 2025-09-07T08:20:06.5466544Z s3-bucket: gha-artifacts 2025-09-07T08:20:06.5466827Z s3-prefix: pytorch/pytorch/17525321311/1/artifact 2025-09-07T08:20:06.5467121Z retention-days: 14 2025-09-07T08:20:06.5467331Z if-no-files-found: ignore 2025-09-07T08:20:06.5467556Z path: debug-*.zip 2025-09-07T08:20:06.5467747Z name: artifact 2025-09-07T08:20:06.5467933Z region: us-east-1 2025-09-07T08:20:06.5468111Z env: 2025-09-07T08:20:06.5468280Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:06.5468675Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:06.5469105Z DEVICE_NAME: 2025-09-07T08:20:06.5469285Z DEVICE_TYPE: 2025-09-07T08:20:06.5469462Z ##[endgroup] 2025-09-07T08:20:06.7995844Z No files were found with the provided path: debug-*.zip. No artifacts will be uploaded. 2025-09-07T08:20:06.8201901Z ##[group]Run # shellcheck disable=SC2156 2025-09-07T08:20:06.8202299Z # shellcheck disable=SC2156 2025-09-07T08:20:06.8202807Z find . -iname "core.[1-9]*" -exec docker exec "${DOCKER_CONTAINER_ID}" sh -c "gdb python {} -ex 'bt' -ex 'q'" \; 2025-09-07T08:20:06.8209518Z shell: /usr/bin/bash -e {0} 2025-09-07T08:20:06.8209735Z env: 2025-09-07T08:20:06.8209913Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:06.8210310Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:06.8210730Z DEVICE_NAME: 2025-09-07T08:20:06.8210909Z DEVICE_TYPE: 2025-09-07T08:20:06.8211087Z ##[endgroup] 2025-09-07T08:20:07.0634754Z Prepare all required actions 2025-09-07T08:20:07.0635130Z Getting action download info 2025-09-07T08:20:07.1694831Z ##[group]Run ./.github/actions/upload-utilization-stats 2025-09-07T08:20:07.1695137Z with: 2025-09-07T08:20:07.1695308Z job_id: 49775354812 2025-09-07T08:20:07.1695806Z job_name: linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal) 2025-09-07T08:20:07.1696398Z workflow_name: inductor-perf-nightly-aarch64 2025-09-07T08:20:07.1696687Z workflow_run_id: 17525321311 2025-09-07T08:20:07.1697053Z workflow_attempt: 1 2025-09-07T08:20:07.1697252Z env: 2025-09-07T08:20:07.1697428Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:07.1697826Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:07.1698247Z DEVICE_NAME: 2025-09-07T08:20:07.1698428Z DEVICE_TYPE: 2025-09-07T08:20:07.1698609Z ##[endgroup] 2025-09-07T08:20:07.1713294Z ##[group]Run echo "workflow_id: 17525321311" 2025-09-07T08:20:07.1713730Z echo "workflow_id: 17525321311" 2025-09-07T08:20:07.1714002Z echo "workflow_attempt: 1" 2025-09-07T08:20:07.1714341Z echo "workflow_Name: inductor-perf-nightly-aarch64" 2025-09-07T08:20:07.1714679Z echo "job_id: 49775354812" 2025-09-07T08:20:07.1715252Z echo "job_name: linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal)" 2025-09-07T08:20:07.1715846Z echo "artifact_prefix: " 2025-09-07T08:20:07.1716095Z python3 --version 2025-09-07T08:20:07.1722177Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:07.1722498Z env: 2025-09-07T08:20:07.1722673Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:07.1723061Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:07.1723481Z DEVICE_NAME: 2025-09-07T08:20:07.1723663Z DEVICE_TYPE: 2025-09-07T08:20:07.1723850Z ##[endgroup] 2025-09-07T08:20:07.1746440Z workflow_id: 17525321311 2025-09-07T08:20:07.1746669Z workflow_attempt: 1 2025-09-07T08:20:07.1746914Z workflow_Name: inductor-perf-nightly-aarch64 2025-09-07T08:20:07.1747196Z job_id: 49775354812 2025-09-07T08:20:07.1747685Z job_name: linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal) 2025-09-07T08:20:07.1748228Z artifact_prefix: 2025-09-07T08:20:07.1757436Z Python 3.9.23 2025-09-07T08:20:07.1790816Z ##[group]Run nick-fields/retry@v3.0.0 2025-09-07T08:20:07.1791069Z with: 2025-09-07T08:20:07.1791239Z shell: bash 2025-09-07T08:20:07.1791418Z timeout_minutes: 5 2025-09-07T08:20:07.1791615Z max_attempts: 5 2025-09-07T08:20:07.1791805Z retry_wait_seconds: 30 2025-09-07T08:20:07.1792273Z command: set -eu python3 -m pip install python-dateutil==2.8.2 boto3==1.35.42 pandas==2.1.3 dataclasses_json==0.6.7 2025-09-07T08:20:07.1792778Z polling_interval_seconds: 1 2025-09-07T08:20:07.1793020Z warning_on_retry: true 2025-09-07T08:20:07.1793245Z continue_on_error: false 2025-09-07T08:20:07.1793450Z env: 2025-09-07T08:20:07.1793617Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:07.1794009Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:07.1794429Z DEVICE_NAME: 2025-09-07T08:20:07.1794608Z DEVICE_TYPE: 2025-09-07T08:20:07.1794786Z ##[endgroup] 2025-09-07T08:20:07.4839875Z Defaulting to user installation because normal site-packages is not writeable 2025-09-07T08:20:07.5487562Z Collecting python-dateutil==2.8.2 2025-09-07T08:20:07.5630410Z Downloading python_dateutil-2.8.2-py2.py3-none-any.whl (247 kB) 2025-09-07T08:20:08.4049688Z Collecting boto3==1.35.42 2025-09-07T08:20:08.4085810Z Downloading boto3-1.35.42-py3-none-any.whl (139 kB) 2025-09-07T08:20:08.8429655Z Collecting pandas==2.1.3 2025-09-07T08:20:08.8485911Z Downloading pandas-2.1.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (14.9 MB) 2025-09-07T08:20:09.0141855Z Collecting dataclasses_json==0.6.7 2025-09-07T08:20:09.0160729Z Using cached dataclasses_json-0.6.7-py3-none-any.whl (28 kB) 2025-09-07T08:20:09.0185638Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil==2.8.2) (1.15.0) 2025-09-07T08:20:09.0228224Z Requirement already satisfied: s3transfer<0.11.0,>=0.10.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from boto3==1.35.42) (0.10.4) 2025-09-07T08:20:09.0232050Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.42) (0.10.0) 2025-09-07T08:20:09.0241604Z Requirement already satisfied: botocore<1.36.0,>=1.35.42 in /home/ec2-user/.local/lib/python3.9/site-packages (from boto3==1.35.42) (1.35.99) 2025-09-07T08:20:09.7692414Z Collecting numpy<2,>=1.22.4 2025-09-07T08:20:09.7728816Z Downloading numpy-1.26.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (14.2 MB) 2025-09-07T08:20:09.9039258Z Collecting tzdata>=2022.1 2025-09-07T08:20:09.9072759Z Downloading tzdata-2025.2-py2.py3-none-any.whl (347 kB) 2025-09-07T08:20:09.9168465Z Requirement already satisfied: pytz>=2020.1 in /usr/lib/python3.9/site-packages (from pandas==2.1.3) (2022.7.1) 2025-09-07T08:20:09.9350781Z Collecting typing-inspect<1,>=0.4.0 2025-09-07T08:20:09.9370862Z Using cached typing_inspect-0.9.0-py3-none-any.whl (8.8 kB) 2025-09-07T08:20:10.0231142Z Collecting marshmallow<4.0.0,>=3.18.0 2025-09-07T08:20:10.0250551Z Using cached marshmallow-3.26.1-py3-none-any.whl (50 kB) 2025-09-07T08:20:10.0314751Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.42->boto3==1.35.42) (1.25.10) 2025-09-07T08:20:10.0762791Z Collecting packaging>=17.0 2025-09-07T08:20:10.0782788Z Using cached packaging-25.0-py3-none-any.whl (66 kB) 2025-09-07T08:20:10.1255337Z Collecting typing-extensions>=3.7.4 2025-09-07T08:20:10.1275321Z Using cached typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-09-07T08:20:10.1412325Z Collecting mypy-extensions>=0.3.0 2025-09-07T08:20:10.1431963Z Using cached mypy_extensions-1.1.0-py3-none-any.whl (5.0 kB) 2025-09-07T08:20:10.3374528Z Installing collected packages: python-dateutil, typing-extensions, packaging, mypy-extensions, tzdata, typing-inspect, numpy, marshmallow, pandas, dataclasses-json, boto3 2025-09-07T08:20:15.3219256Z Attempting uninstall: boto3 2025-09-07T08:20:15.3220752Z Found existing installation: boto3 1.35.33 2025-09-07T08:20:15.3310202Z Uninstalling boto3-1.35.33: 2025-09-07T08:20:15.3323099Z Successfully uninstalled boto3-1.35.33 2025-09-07T08:20:15.3845315Z Successfully installed boto3-1.35.42 dataclasses-json-0.6.7 marshmallow-3.26.1 mypy-extensions-1.1.0 numpy-1.26.4 packaging-25.0 pandas-2.1.3 python-dateutil-2.8.2 typing-extensions-4.15.0 typing-inspect-0.9.0 tzdata-2025.2 2025-09-07T08:20:16.2550882Z Command completed after 1 attempt(s). 2025-09-07T08:20:16.2600843Z ##[group]Run python3 -m tools.stats.upload_utilization_stats.upload_utilization_stats \ 2025-09-07T08:20:16.2601478Z python3 -m tools.stats.upload_utilization_stats.upload_utilization_stats \ 2025-09-07T08:20:16.2601912Z  --workflow-run-id "17525321311" \ 2025-09-07T08:20:16.2602261Z  --workflow-name "inductor-perf-nightly-aarch64" \ 2025-09-07T08:20:16.2602607Z  --workflow-run-attempt "1" \ 2025-09-07T08:20:16.2602877Z  --job-id "49775354812" \ 2025-09-07T08:20:16.2603449Z  --job-name "linux-jammy-aarch64-py3.10-inductor / test (inductor_torchbench_perf_cpu_aarch64, 8, 15, linux.arm64.m7g.metal)" \ 2025-09-07T08:20:16.2604037Z  --local-path "" \ 2025-09-07T08:20:16.2604277Z  --artifact-prefix "" 2025-09-07T08:20:16.2610816Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:16.2611135Z env: 2025-09-07T08:20:16.2611309Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:16.2611710Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:16.2612367Z DEVICE_NAME: 2025-09-07T08:20:16.2612572Z DEVICE_TYPE: 2025-09-07T08:20:16.2612760Z ##[endgroup] 2025-09-07T08:20:16.8261842Z repo: pytorch/pytorch 2025-09-07T08:20:16.8262148Z Search for test log in s3 bucket: ossci-utilization 2025-09-07T08:20:16.8262704Z Downloading logs-test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812.zip 2025-09-07T08:20:16.8263508Z extracting usage_log.txt from zip file logs-test-inductor_torchbench_perf_cpu_aarch64-8-15-linux.arm64.m7g.metal_49775354812.zip 2025-09-07T08:20:16.8264353Z ::warning trying to download test log failed by: "There is no item named 'usage_log.txt' in the archive" 2025-09-07T08:20:16.8265153Z [Log Model] Failed to process test log, metadata is None 2025-09-07T08:20:16.9039347Z ##[group]Run pytorch/test-infra/.github/actions/teardown-linux@main 2025-09-07T08:20:16.9039795Z with: 2025-09-07T08:20:16.9039957Z env: 2025-09-07T08:20:16.9040126Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:16.9040531Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:16.9041094Z DEVICE_NAME: 2025-09-07T08:20:16.9041279Z DEVICE_TYPE: 2025-09-07T08:20:16.9041455Z ##[endgroup] 2025-09-07T08:20:16.9057083Z ##[group]Run set -eou pipefail 2025-09-07T08:20:16.9057354Z set -eou pipefail 2025-09-07T08:20:16.9057573Z  2025-09-07T08:20:16.9057879Z echo "Holding runner for 2 hours until all ssh sessions have logged out" 2025-09-07T08:20:16.9058264Z for _ in $(seq 1440); do 2025-09-07T08:20:16.9058543Z  # Break if no ssh session exists anymore 2025-09-07T08:20:16.9058848Z  if [ "$(who)" = "" ]; then 2025-09-07T08:20:16.9059093Z  break 2025-09-07T08:20:16.9059304Z  fi 2025-09-07T08:20:16.9059488Z  echo "." 2025-09-07T08:20:16.9059682Z  sleep 5 2025-09-07T08:20:16.9059872Z done 2025-09-07T08:20:16.9066638Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:16.9066981Z env: 2025-09-07T08:20:16.9067151Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:16.9067545Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:16.9067966Z DEVICE_NAME: 2025-09-07T08:20:16.9068146Z DEVICE_TYPE: 2025-09-07T08:20:16.9068320Z ##[endgroup] 2025-09-07T08:20:16.9089889Z Holding runner for 2 hours until all ssh sessions have logged out 2025-09-07T08:20:16.9183406Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-09-07T08:20:16.9183969Z # ignore expansion of "docker ps -q" since it could be empty 2025-09-07T08:20:16.9184333Z # shellcheck disable=SC2046 2025-09-07T08:20:16.9184626Z docker stop $(docker ps -q) || true 2025-09-07T08:20:16.9184916Z # Prune all of the docker images 2025-09-07T08:20:16.9185190Z docker system prune -af 2025-09-07T08:20:16.9190904Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:16.9191237Z env: 2025-09-07T08:20:16.9191413Z GIT_DEFAULT_BRANCH: main 2025-09-07T08:20:16.9191807Z DOCKER_CONTAINER_ID: 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:16.9192227Z DEVICE_NAME: 2025-09-07T08:20:16.9192410Z DEVICE_TYPE: 2025-09-07T08:20:16.9192591Z ##[endgroup] 2025-09-07T08:20:27.8518815Z 4646a3beb296 2025-09-07T08:20:28.4319325Z Deleted Containers: 2025-09-07T08:20:28.4319693Z 4646a3beb29629d5dbe2c669fa4484f37bbfd52b1c7931e2d47dbfc4136ac79d 2025-09-07T08:20:28.4319991Z 2025-09-07T08:20:35.1368232Z Deleted Images: 2025-09-07T08:20:35.1369097Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-aarch64-py3.10-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T08:20:35.1370297Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image@sha256:c8bf3aa0f88e5662ff72435fb728fa2432ad59709b953fd68e95432215dfee4a 2025-09-07T08:20:35.1371081Z deleted: sha256:2e41fa7f17d613e4abb96e4b9f7f4a2b2a08134df169ebcbcbc5a819408dc5b3 2025-09-07T08:20:35.1371653Z deleted: sha256:6652ecf65d8c4cd0c3717dc16a11439cde78848e05f81f231f3a87c2bdca3e31 2025-09-07T08:20:35.1372193Z deleted: sha256:cdacb207081fe2073693646a77441b430d48e78fc30b78acab75e11e4760c572 2025-09-07T08:20:35.1372740Z deleted: sha256:683e8048cce4d643d0f939a1f8c9c5c3c0ce002dc48c0fa451fe18d0b1c48de3 2025-09-07T08:20:35.1373291Z deleted: sha256:0fddbb2f18c4a48c5461b156b143d31c4ca8107a6edb9b402c3a39c0bf9fe568 2025-09-07T08:20:35.1373836Z deleted: sha256:2987a1638cc0929b9a40cbca7555efd1789f257f6c333fa150cf8b4327b77dae 2025-09-07T08:20:35.1374681Z deleted: sha256:41137f5400c89e43fca7c551fd54168da73945dac9fdf0afaacae72c19643b23 2025-09-07T08:20:35.1375455Z deleted: sha256:e9f27082d8cd15d6ca1e83107dae4d81aa66b2776405abc976b5b45b26e5b614 2025-09-07T08:20:35.1376014Z deleted: sha256:432d0f6988aeacde28cb94fb99157a15df2c66f24c43c0e9bde3a0a9c7f5fd5e 2025-09-07T08:20:35.1376567Z deleted: sha256:e68c88a1c6e17ba9b35198e38432d7935df0a7f7ba14a6c7227aa6916db22829 2025-09-07T08:20:35.1377212Z deleted: sha256:0648ee6835ae7465133c4d0b5a5f8db3396449cc87e379c4d08c3cf0f0d72790 2025-09-07T08:20:35.1377755Z deleted: sha256:c788b3fc42b63a03b0b0dcbe9cb3e175fd20f36cce72c96c982b961976633e1a 2025-09-07T08:20:35.1378308Z deleted: sha256:0cfdce690c6f25c1d20cc7a4122f963edd5a9d293971bd557816f8df973cae9a 2025-09-07T08:20:35.1378869Z deleted: sha256:8b2cf8dabda0f248b5f3951afab3dfe31ebe0b9aacc7653450feafa069b0fc5c 2025-09-07T08:20:35.1379431Z deleted: sha256:7aa8e1c892c0a644d96473f8b25a5fbfcfbf05fff833945a6fc260be69904aaf 2025-09-07T08:20:35.1379975Z deleted: sha256:2c09938ca84e0e679f691e7d91030f4058669af9e6fd7027829e4d8531bc8348 2025-09-07T08:20:35.1380508Z deleted: sha256:4099e202c582c269415ef11b82e5365240175d3114ca6716debf0095f7872fe0 2025-09-07T08:20:35.1381086Z deleted: sha256:6c6d4515fc408ad9c164ff0965d6ff336da9365e5b18f76ada165029bbc36887 2025-09-07T08:20:35.1381623Z deleted: sha256:148299a2652e6d627dded619e600c322e4fb39645cc19030c4ddae12f61bad96 2025-09-07T08:20:35.1382172Z deleted: sha256:69eddf8f3e9aec34aa0a3f2a89093416214323d4adbe9b1aed4b0558fee323fb 2025-09-07T08:20:35.1382716Z deleted: sha256:6968d778217eb38761b38a2d729b5797adcd447ae2c8b86f086500cf9acda941 2025-09-07T08:20:35.1383257Z deleted: sha256:fab1ae72d55eb09f8ae4f65c6fb33a03b04d7444ba8bb67f50f8d79a7f542ccd 2025-09-07T08:20:35.1383818Z deleted: sha256:a1fda6dbc5bfe6b668e8c9d45050f22831d449e62fb8faa8959c54ccb76a2a19 2025-09-07T08:20:35.1384375Z deleted: sha256:f94ec736dca563c050d42cb136a1d077f0a3009bafa3124b1e57db9bd87caf98 2025-09-07T08:20:35.1384919Z deleted: sha256:3d6fa0f32b9e33e807f102cb369be37b880607f887f234713fc976a5989f8c7c 2025-09-07T08:20:35.1385454Z deleted: sha256:7b1d883b44c717721d70015acf29d4520c78f0b6504992b47f6413dc8597a87c 2025-09-07T08:20:35.1385992Z deleted: sha256:153a678e92c2c9920035ba75ed2204627fe511bb436233bd2ed3f7550e52f8f8 2025-09-07T08:20:35.1386522Z deleted: sha256:5809218e6eef2955507398b7085769fc8243dcc6b0dfbb0a12d0625cc58d76d7 2025-09-07T08:20:35.1387065Z deleted: sha256:1ff5ea88cfdb2bd3c11a31fa039901f66418784b7bf80c5e1ab1005d6759d965 2025-09-07T08:20:35.1387613Z deleted: sha256:adff4f2c3e54d999e31f701841d20fc97d2d2e8625409482e2f68d56a5d6f1f9 2025-09-07T08:20:35.1388150Z deleted: sha256:b73bf58e2c50d0952db7196610a0c2e1cf8e80b0f0d6ee5983820d739007172a 2025-09-07T08:20:35.1388711Z deleted: sha256:f9fe16b2ac2c8d08d5edb5aaec0fdc7679dcce46edac6f1c1ffdb0396432435c 2025-09-07T08:20:35.1389271Z deleted: sha256:ccb800bdac670193c46df2566ad0bba3276961927c2289bb26cf6a50a9941a5a 2025-09-07T08:20:35.1389830Z deleted: sha256:efcc3ba597db01da6ab64f17cab91dca665a6b38aad59a3ac25ad2207190e2ad 2025-09-07T08:20:35.1390391Z deleted: sha256:43925a12249bbf4a9e7baaaaab003d97834bd835de321c4e85c06c72be244d3e 2025-09-07T08:20:35.1390938Z deleted: sha256:48224c5ac9c586c2b4fd0dae773465e7863835df439e0d06f5f5f33b942b53c7 2025-09-07T08:20:35.1391480Z deleted: sha256:cd8a667a7367826e5a71788a5fc206aa6326c009b8e6ec1fa854ee5bcd8a4ab2 2025-09-07T08:20:35.1392028Z deleted: sha256:1f5c95653ed17031bdebb1760afd7037ac22a79e2f1d03fbf63a66acab05283e 2025-09-07T08:20:35.1392575Z deleted: sha256:65a0060afbafbe1be2b064810166aad5904a653e28fe2d6f1650f880602b3a05 2025-09-07T08:20:35.1393128Z deleted: sha256:fc28ebdabc3e1b402c370ed2f2c9c1454cea0d0bfd5dc17a9c4e510334b783bd 2025-09-07T08:20:35.1393677Z deleted: sha256:3323792295637149f5f7ea7dd25ec9fe67e037332ae0b5f623f748f1c96c59ea 2025-09-07T08:20:35.1394218Z deleted: sha256:b295beb9947c94c8bc658c0a8c4e1b917e82ab4d3ee262cc3fa66c2307f6ed60 2025-09-07T08:20:35.1394769Z deleted: sha256:5208a565c4d2f24fc68cd841b1dedc82e5a74bb6bdf2788560c935cbc7a22fa2 2025-09-07T08:20:35.1395416Z deleted: sha256:1d3f4621c21190da54e6c3768eccf2c1fc89ac85e7c8805fc7ddbeadbfe774c0 2025-09-07T08:20:35.1396119Z deleted: sha256:62429aac645474fcea0831d223a2de22ed52a42d89545cf8e752eaed785b21aa 2025-09-07T08:20:35.1396667Z deleted: sha256:a8ba38ecf53f39010e5a5619b4579989bd348a425eeac2ac1ac805cc7939c654 2025-09-07T08:20:35.1397208Z deleted: sha256:f629d5c15ccf7978129243e698404040c27a4bb120c85ac1d5ba016939d2ce05 2025-09-07T08:20:35.1397779Z deleted: sha256:9a7886837f4dd34f8d5846660f67c96dc24e21446d7f60c7afcdac7490cad747 2025-09-07T08:20:35.1398321Z deleted: sha256:f719e81bc5cdd2bdd64bb3b5277f3f8cfff7bf8253786100f36018839899f1e8 2025-09-07T08:20:35.1398859Z deleted: sha256:a24d59924161fe0d98958bd15e633af9b4d258b733c55183a24420331a3e80cd 2025-09-07T08:20:35.1399399Z deleted: sha256:1ddcf77911d3f097ef30f69d14d875629aa1f181b377aad41b42de2fcb596b9a 2025-09-07T08:20:35.1399939Z deleted: sha256:1fc66fd0229379e6e73211372e8fea890fe92b68a4f1c59e4048c11179736901 2025-09-07T08:20:35.1400476Z deleted: sha256:a106c2479245f0c642a475e3cfa340d01d9d4db9ac812a96ecf7c91670f7dba9 2025-09-07T08:20:35.1401030Z deleted: sha256:0d4605a7d267c7a3f40abff0d6feeaf18c42b29e8fc7dc195b7d5ff51d61ba16 2025-09-07T08:20:35.1401585Z deleted: sha256:3a5d548af5eeb022bee4704f3d07d82cfb120533a458f465687060de99d22cab 2025-09-07T08:20:35.1402136Z deleted: sha256:1ba4dd55bbff0b1483d5aa6ac69455c059fd55440d0ef107f04a1a3fb880ed0d 2025-09-07T08:20:35.1402680Z deleted: sha256:7c823991843bbb307f299c8870d0fb1215eb054e3766485fd70f439834af971c 2025-09-07T08:20:35.1403210Z deleted: sha256:5a423716350718e4bb2581561bb6e72148aa81d3f3918ee1ac32d29abf25f8f9 2025-09-07T08:20:35.1403740Z deleted: sha256:633ead0f9f6b4e781222bea3bf4373b3a9c044009152774f939f94865a9ead4f 2025-09-07T08:20:35.1404280Z deleted: sha256:e76f65682ca4eb9ff285136c90f964e2ffd7a23c131f50350695260c7e094628 2025-09-07T08:20:35.1404818Z deleted: sha256:31284cfa98ff07902046399975b4f8b62bd6638b64d7aad8c5b91b3d35a1a2d3 2025-09-07T08:20:35.1405352Z deleted: sha256:f9f7cb49420cf2f672d192d0fff319bdc216b26604a4a2232383f5be3546f926 2025-09-07T08:20:35.1405899Z deleted: sha256:a3d4f3167a8b8f31554318b561cbeb99e6e44a8bfffc1bb345b7b8b481ac87e4 2025-09-07T08:20:35.1406458Z deleted: sha256:b8ed795c3ce4f35debb878f1b493fe5e273943defebed508c8bd438d38b9ece8 2025-09-07T08:20:35.1407008Z deleted: sha256:8e57222e830ea263d9fac3a7cb42f7d478e5fc33573f10cf90146258aac5ad7f 2025-09-07T08:20:35.1407550Z deleted: sha256:dafd40996908c1879f69cc4aab5d3575ccf1c98f3849686ef476356b2617bac8 2025-09-07T08:20:35.1408091Z deleted: sha256:53d43a4979316ffdaee3222028137b624137a7ce671d2e7530d6e6e7cb2e8a3d 2025-09-07T08:20:35.1408632Z deleted: sha256:0b3e28d0c91a135e80d39d356abd67ee0d4dc88a318adf2efbd2113d21dd4b76 2025-09-07T08:20:35.1409188Z deleted: sha256:b533355bcf7b799bd41ad303a8674e0bbe9aedb6ce67965b2afbebb552afad0a 2025-09-07T08:20:35.1409757Z deleted: sha256:a8fabe19faafe6ffecf1f9b2ec6a05e39ddc6c1f6ed92b1ce025babb95bfe559 2025-09-07T08:20:35.1410372Z deleted: sha256:dc0d79daf81dbaf7349b2321094d7ceda164286c3413a287a24ca588ee4eb70e 2025-09-07T08:20:35.1410931Z deleted: sha256:f1fd9640001c0650983f7cd3dfc7b30ad931b21b6630e21f1aa4afe9420d6451 2025-09-07T08:20:35.1411480Z deleted: sha256:e9579c57c0041bc8e14b0f43f19d42fd24dd1f2058e5eaf7bcb2bf30b8489901 2025-09-07T08:20:35.1412029Z deleted: sha256:4f2e0e0d8c0b855b156b25ad59d5b37da660b41b8d33585bb8e63e7e8cf9e45b 2025-09-07T08:20:35.1412575Z deleted: sha256:cf5839f5e5f7f96677a3da3b897e6e7103684ef845ca93e71b10b1319b702900 2025-09-07T08:20:35.1413120Z deleted: sha256:1070d402a8b706082cd62aa44b1878e26b9de98e2ede72e6689f781e5bcc2def 2025-09-07T08:20:35.1413666Z deleted: sha256:3e04ae02e3b48a52cd15f8595efd67e8ab7b3e9766f7d1e12a25c8c76f3a86bf 2025-09-07T08:20:35.1413996Z 2025-09-07T08:20:35.1414095Z Total reclaimed space: 30.63GB 2025-09-07T08:20:35.1509235Z Post job cleanup. 2025-09-07T08:20:35.1548554Z Post job cleanup. 2025-09-07T08:20:35.2412497Z [command]/usr/bin/git version 2025-09-07T08:20:35.2452154Z git version 2.47.1 2025-09-07T08:20:35.2486913Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/c87b0f1c-bb39-4e08-9854-9f6537afdd5d/.gitconfig' 2025-09-07T08:20:35.2495973Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/c87b0f1c-bb39-4e08-9854-9f6537afdd5d' before making global git config changes 2025-09-07T08:20:35.2497027Z Adding repository directory to the temporary git global config as a safe directory 2025-09-07T08:20:35.2501516Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-09-07T08:20:35.2538956Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-09-07T08:20:35.2577661Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-09-07T08:20:35.2861374Z Entering 'android/libs/fbjni' 2025-09-07T08:20:35.2917321Z Entering 'third_party/FP16' 2025-09-07T08:20:35.2972960Z Entering 'third_party/FXdiv' 2025-09-07T08:20:35.3028951Z Entering 'third_party/NNPACK' 2025-09-07T08:20:35.3084847Z Entering 'third_party/NVTX' 2025-09-07T08:20:35.3139265Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T08:20:35.3193788Z Entering 'third_party/XNNPACK' 2025-09-07T08:20:35.3258009Z Entering 'third_party/aiter' 2025-09-07T08:20:35.3312586Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T08:20:35.3372070Z Entering 'third_party/benchmark' 2025-09-07T08:20:35.3426097Z Entering 'third_party/composable_kernel' 2025-09-07T08:20:35.3486315Z Entering 'third_party/cpp-httplib' 2025-09-07T08:20:35.3540256Z Entering 'third_party/cpuinfo' 2025-09-07T08:20:35.3594525Z Entering 'third_party/cudnn_frontend' 2025-09-07T08:20:35.3648216Z Entering 'third_party/cutlass' 2025-09-07T08:20:35.3708330Z Entering 'third_party/fbgemm' 2025-09-07T08:20:35.3762832Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T08:20:35.3814192Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T08:20:35.3871423Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T08:20:35.3923663Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T08:20:35.3982396Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T08:20:35.4033431Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T08:20:35.4084274Z Entering 'third_party/fbgemm/external/json' 2025-09-07T08:20:35.4139528Z Entering 'third_party/flash-attention' 2025-09-07T08:20:35.4192981Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T08:20:35.4249735Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T08:20:35.4308970Z Entering 'third_party/flatbuffers' 2025-09-07T08:20:35.4364819Z Entering 'third_party/fmt' 2025-09-07T08:20:35.4418265Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T08:20:35.4472086Z Entering 'third_party/gloo' 2025-09-07T08:20:35.4524529Z Entering 'third_party/googletest' 2025-09-07T08:20:35.4577325Z Entering 'third_party/ideep' 2025-09-07T08:20:35.4629691Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T08:20:35.4687481Z Entering 'third_party/ittapi' 2025-09-07T08:20:35.4740807Z Entering 'third_party/kineto' 2025-09-07T08:20:35.4793890Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T08:20:35.4844693Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T08:20:35.4897056Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T08:20:35.4949387Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T08:20:35.5000624Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T08:20:35.5051683Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T08:20:35.5105852Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T08:20:35.5158309Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T08:20:35.5209730Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T08:20:35.5261522Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T08:20:35.5314542Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T08:20:35.5366100Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T08:20:35.5419002Z Entering 'third_party/kleidiai' 2025-09-07T08:20:35.5473115Z Entering 'third_party/mimalloc' 2025-09-07T08:20:35.5527142Z Entering 'third_party/nlohmann' 2025-09-07T08:20:35.5581648Z Entering 'third_party/onnx' 2025-09-07T08:20:35.5647357Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T08:20:35.5702995Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T08:20:35.5756834Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T08:20:35.5809185Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T08:20:35.5861069Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T08:20:35.5912917Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T08:20:35.5965041Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T08:20:35.6016528Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T08:20:35.6067256Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T08:20:35.6117480Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T08:20:35.6171144Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T08:20:35.6224734Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T08:20:35.6292752Z Entering 'third_party/pocketfft' 2025-09-07T08:20:35.6345124Z Entering 'third_party/protobuf' 2025-09-07T08:20:35.6401684Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T08:20:35.6453453Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T08:20:35.6506776Z Entering 'third_party/psimd' 2025-09-07T08:20:35.6560222Z Entering 'third_party/pthreadpool' 2025-09-07T08:20:35.6614703Z Entering 'third_party/pybind11' 2025-09-07T08:20:35.6668214Z Entering 'third_party/python-peachpy' 2025-09-07T08:20:35.6722305Z Entering 'third_party/sleef' 2025-09-07T08:20:35.6774908Z Entering 'third_party/tensorpipe' 2025-09-07T08:20:35.6827306Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T08:20:35.6877902Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T08:20:35.6928753Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T08:20:35.6980848Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T08:20:35.7030504Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T08:20:35.7105010Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-09-07T08:20:35.7125257Z http.https://github.com/.extraheader 2025-09-07T08:20:35.7135346Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2025-09-07T08:20:35.7165452Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-09-07T08:20:35.7440655Z Entering 'android/libs/fbjni' 2025-09-07T08:20:35.7476023Z http.https://github.com/.extraheader 2025-09-07T08:20:35.7509730Z Entering 'third_party/FP16' 2025-09-07T08:20:35.7545013Z http.https://github.com/.extraheader 2025-09-07T08:20:35.7578876Z Entering 'third_party/FXdiv' 2025-09-07T08:20:35.7613980Z http.https://github.com/.extraheader 2025-09-07T08:20:35.7648718Z Entering 'third_party/NNPACK' 2025-09-07T08:20:35.7683879Z http.https://github.com/.extraheader 2025-09-07T08:20:35.7717097Z Entering 'third_party/NVTX' 2025-09-07T08:20:35.7752100Z http.https://github.com/.extraheader 2025-09-07T08:20:35.7786313Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T08:20:35.7821407Z http.https://github.com/.extraheader 2025-09-07T08:20:35.7854602Z Entering 'third_party/XNNPACK' 2025-09-07T08:20:35.7890308Z http.https://github.com/.extraheader 2025-09-07T08:20:35.7932390Z Entering 'third_party/aiter' 2025-09-07T08:20:35.7967411Z http.https://github.com/.extraheader 2025-09-07T08:20:35.8000871Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T08:20:35.8035350Z http.https://github.com/.extraheader 2025-09-07T08:20:35.8075511Z Entering 'third_party/benchmark' 2025-09-07T08:20:35.8110559Z http.https://github.com/.extraheader 2025-09-07T08:20:35.8144305Z Entering 'third_party/composable_kernel' 2025-09-07T08:20:35.8179176Z http.https://github.com/.extraheader 2025-09-07T08:20:35.8218432Z Entering 'third_party/cpp-httplib' 2025-09-07T08:20:35.8253726Z http.https://github.com/.extraheader 2025-09-07T08:20:35.8287348Z Entering 'third_party/cpuinfo' 2025-09-07T08:20:35.8322480Z http.https://github.com/.extraheader 2025-09-07T08:20:35.8355553Z Entering 'third_party/cudnn_frontend' 2025-09-07T08:20:35.8390988Z http.https://github.com/.extraheader 2025-09-07T08:20:35.8424419Z Entering 'third_party/cutlass' 2025-09-07T08:20:35.8459613Z http.https://github.com/.extraheader 2025-09-07T08:20:35.8498821Z Entering 'third_party/fbgemm' 2025-09-07T08:20:35.8534137Z http.https://github.com/.extraheader 2025-09-07T08:20:35.8568892Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T08:20:35.8603000Z http.https://github.com/.extraheader 2025-09-07T08:20:35.8636123Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T08:20:35.8670948Z http.https://github.com/.extraheader 2025-09-07T08:20:35.8709109Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T08:20:35.8743824Z http.https://github.com/.extraheader 2025-09-07T08:20:35.8776417Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T08:20:35.8810967Z http.https://github.com/.extraheader 2025-09-07T08:20:35.8849731Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T08:20:35.8884214Z http.https://github.com/.extraheader 2025-09-07T08:20:35.8916922Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T08:20:35.8951450Z http.https://github.com/.extraheader 2025-09-07T08:20:35.8983226Z Entering 'third_party/fbgemm/external/json' 2025-09-07T08:20:35.9017386Z http.https://github.com/.extraheader 2025-09-07T08:20:35.9052817Z Entering 'third_party/flash-attention' 2025-09-07T08:20:35.9087832Z http.https://github.com/.extraheader 2025-09-07T08:20:35.9120525Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T08:20:35.9154363Z http.https://github.com/.extraheader 2025-09-07T08:20:35.9190882Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T08:20:35.9225130Z http.https://github.com/.extraheader 2025-09-07T08:20:35.9266519Z Entering 'third_party/flatbuffers' 2025-09-07T08:20:35.9302351Z http.https://github.com/.extraheader 2025-09-07T08:20:35.9336793Z Entering 'third_party/fmt' 2025-09-07T08:20:35.9372419Z http.https://github.com/.extraheader 2025-09-07T08:20:35.9404934Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T08:20:35.9440108Z http.https://github.com/.extraheader 2025-09-07T08:20:35.9473354Z Entering 'third_party/gloo' 2025-09-07T08:20:35.9508282Z http.https://github.com/.extraheader 2025-09-07T08:20:35.9542021Z Entering 'third_party/googletest' 2025-09-07T08:20:35.9576806Z http.https://github.com/.extraheader 2025-09-07T08:20:35.9610115Z Entering 'third_party/ideep' 2025-09-07T08:20:35.9645250Z http.https://github.com/.extraheader 2025-09-07T08:20:35.9677935Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T08:20:35.9712295Z http.https://github.com/.extraheader 2025-09-07T08:20:35.9752296Z Entering 'third_party/ittapi' 2025-09-07T08:20:35.9788833Z http.https://github.com/.extraheader 2025-09-07T08:20:35.9822122Z Entering 'third_party/kineto' 2025-09-07T08:20:35.9857099Z http.https://github.com/.extraheader 2025-09-07T08:20:35.9890157Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T08:20:35.9924112Z http.https://github.com/.extraheader 2025-09-07T08:20:35.9957159Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T08:20:35.9991838Z http.https://github.com/.extraheader 2025-09-07T08:20:36.0026084Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T08:20:36.0060674Z http.https://github.com/.extraheader 2025-09-07T08:20:36.0094463Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T08:20:36.0128941Z http.https://github.com/.extraheader 2025-09-07T08:20:36.0161467Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T08:20:36.0196587Z http.https://github.com/.extraheader 2025-09-07T08:20:36.0228711Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T08:20:36.0263511Z http.https://github.com/.extraheader 2025-09-07T08:20:36.0298540Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T08:20:36.0333096Z http.https://github.com/.extraheader 2025-09-07T08:20:36.0366157Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T08:20:36.0399347Z http.https://github.com/.extraheader 2025-09-07T08:20:36.0432068Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T08:20:36.0466304Z http.https://github.com/.extraheader 2025-09-07T08:20:36.0500830Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T08:20:36.0535000Z http.https://github.com/.extraheader 2025-09-07T08:20:36.0569989Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T08:20:36.0604046Z http.https://github.com/.extraheader 2025-09-07T08:20:36.0636221Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T08:20:36.0670420Z http.https://github.com/.extraheader 2025-09-07T08:20:36.0704091Z Entering 'third_party/kleidiai' 2025-09-07T08:20:36.0739806Z http.https://github.com/.extraheader 2025-09-07T08:20:36.0773463Z Entering 'third_party/mimalloc' 2025-09-07T08:20:36.0808259Z http.https://github.com/.extraheader 2025-09-07T08:20:36.0841333Z Entering 'third_party/nlohmann' 2025-09-07T08:20:36.0876500Z http.https://github.com/.extraheader 2025-09-07T08:20:36.0911352Z Entering 'third_party/onnx' 2025-09-07T08:20:36.0946543Z http.https://github.com/.extraheader 2025-09-07T08:20:36.0991479Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T08:20:36.1025946Z http.https://github.com/.extraheader 2025-09-07T08:20:36.1061517Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T08:20:36.1096798Z http.https://github.com/.extraheader 2025-09-07T08:20:36.1131491Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T08:20:36.1166439Z http.https://github.com/.extraheader 2025-09-07T08:20:36.1198798Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T08:20:36.1232496Z http.https://github.com/.extraheader 2025-09-07T08:20:36.1265019Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T08:20:36.1298924Z http.https://github.com/.extraheader 2025-09-07T08:20:36.1330831Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T08:20:36.1364887Z http.https://github.com/.extraheader 2025-09-07T08:20:36.1398645Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T08:20:36.1432635Z http.https://github.com/.extraheader 2025-09-07T08:20:36.1463999Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T08:20:36.1498092Z http.https://github.com/.extraheader 2025-09-07T08:20:36.1531035Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T08:20:36.1564980Z http.https://github.com/.extraheader 2025-09-07T08:20:36.1597213Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T08:20:36.1631369Z http.https://github.com/.extraheader 2025-09-07T08:20:36.1666161Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T08:20:36.1700091Z http.https://github.com/.extraheader 2025-09-07T08:20:36.1734521Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T08:20:36.1767844Z http.https://github.com/.extraheader 2025-09-07T08:20:36.1815656Z Entering 'third_party/pocketfft' 2025-09-07T08:20:36.1851046Z http.https://github.com/.extraheader 2025-09-07T08:20:36.1884788Z Entering 'third_party/protobuf' 2025-09-07T08:20:36.1920422Z http.https://github.com/.extraheader 2025-09-07T08:20:36.1955598Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T08:20:36.1990166Z http.https://github.com/.extraheader 2025-09-07T08:20:36.2022335Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T08:20:36.2057076Z http.https://github.com/.extraheader 2025-09-07T08:20:36.2091808Z Entering 'third_party/psimd' 2025-09-07T08:20:36.2127806Z http.https://github.com/.extraheader 2025-09-07T08:20:36.2162056Z Entering 'third_party/pthreadpool' 2025-09-07T08:20:36.2197438Z http.https://github.com/.extraheader 2025-09-07T08:20:36.2231808Z Entering 'third_party/pybind11' 2025-09-07T08:20:36.2266689Z http.https://github.com/.extraheader 2025-09-07T08:20:36.2299305Z Entering 'third_party/python-peachpy' 2025-09-07T08:20:36.2334402Z http.https://github.com/.extraheader 2025-09-07T08:20:36.2367121Z Entering 'third_party/sleef' 2025-09-07T08:20:36.2402614Z http.https://github.com/.extraheader 2025-09-07T08:20:36.2436165Z Entering 'third_party/tensorpipe' 2025-09-07T08:20:36.2471284Z http.https://github.com/.extraheader 2025-09-07T08:20:36.2503901Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T08:20:36.2538670Z http.https://github.com/.extraheader 2025-09-07T08:20:36.2570865Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T08:20:36.2604592Z http.https://github.com/.extraheader 2025-09-07T08:20:36.2636821Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T08:20:36.2671415Z http.https://github.com/.extraheader 2025-09-07T08:20:36.2704082Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T08:20:36.2738064Z http.https://github.com/.extraheader 2025-09-07T08:20:36.2770385Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T08:20:36.2804986Z http.https://github.com/.extraheader 2025-09-07T08:20:36.2934766Z A job completed hook has been configured by the self-hosted runner administrator 2025-09-07T08:20:36.2951782Z ##[group]Run '/home/ec2-user/runner-scripts/after_job.sh' 2025-09-07T08:20:36.2957210Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T08:20:36.2957552Z ##[endgroup] 2025-09-07T08:20:36.3037111Z [!ALERT!] Swap in detected! [!ALERT!] 2025-09-07T08:20:46.0956888Z [!ALERT!] Swap out detected [!ALERT!] 2025-09-07T08:21:02.6495543Z Cleaning up orphan processes